var/home/core/zuul-output/0000755000175000017500000000000015111356230014522 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111363212015464 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003764421215111363203017701 0ustar rootrootNov 25 16:46:55 crc systemd[1]: Starting Kubernetes Kubelet... Nov 25 16:46:55 crc restorecon[4706]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:55 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 16:46:56 crc restorecon[4706]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 16:46:56 crc restorecon[4706]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 25 16:46:57 crc kubenswrapper[4802]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 16:46:57 crc kubenswrapper[4802]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 25 16:46:57 crc kubenswrapper[4802]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 16:46:57 crc kubenswrapper[4802]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 16:46:57 crc kubenswrapper[4802]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 25 16:46:57 crc kubenswrapper[4802]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.302773 4802 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309095 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309169 4802 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309181 4802 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309190 4802 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309198 4802 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309206 4802 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309216 4802 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309225 4802 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309246 4802 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309255 4802 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309263 4802 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309272 4802 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309280 4802 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309289 4802 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309298 4802 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309306 4802 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309314 4802 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309323 4802 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309332 4802 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309340 4802 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309349 4802 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309357 4802 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309365 4802 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309374 4802 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309382 4802 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309391 4802 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309399 4802 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309408 4802 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309416 4802 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309424 4802 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309436 4802 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309446 4802 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309456 4802 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309466 4802 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309475 4802 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309484 4802 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309492 4802 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309502 4802 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309511 4802 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309520 4802 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309529 4802 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309551 4802 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309561 4802 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309570 4802 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309580 4802 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309589 4802 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309599 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309608 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309617 4802 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309625 4802 feature_gate.go:330] unrecognized feature gate: Example Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309634 4802 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309643 4802 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309651 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309660 4802 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309669 4802 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309680 4802 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309690 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309699 4802 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309707 4802 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309715 4802 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309724 4802 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309732 4802 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309744 4802 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309755 4802 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309764 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309772 4802 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309781 4802 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309791 4802 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309800 4802 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309809 4802 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.309818 4802 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.310882 4802 flags.go:64] FLAG: --address="0.0.0.0" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.310906 4802 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.310923 4802 flags.go:64] FLAG: --anonymous-auth="true" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.310934 4802 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.310943 4802 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.310950 4802 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.310967 4802 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.310976 4802 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.310984 4802 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.310991 4802 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.310999 4802 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311007 4802 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311015 4802 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311024 4802 flags.go:64] FLAG: --cgroup-root="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311031 4802 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311038 4802 flags.go:64] FLAG: --client-ca-file="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311045 4802 flags.go:64] FLAG: --cloud-config="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311052 4802 flags.go:64] FLAG: --cloud-provider="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311059 4802 flags.go:64] FLAG: --cluster-dns="[]" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311068 4802 flags.go:64] FLAG: --cluster-domain="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311075 4802 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311082 4802 flags.go:64] FLAG: --config-dir="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311090 4802 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311097 4802 flags.go:64] FLAG: --container-log-max-files="5" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311107 4802 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311114 4802 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311146 4802 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311156 4802 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311164 4802 flags.go:64] FLAG: --contention-profiling="false" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311172 4802 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311180 4802 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311187 4802 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311194 4802 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311203 4802 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311211 4802 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311218 4802 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311225 4802 flags.go:64] FLAG: --enable-load-reader="false" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311232 4802 flags.go:64] FLAG: --enable-server="true" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311242 4802 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311253 4802 flags.go:64] FLAG: --event-burst="100" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311261 4802 flags.go:64] FLAG: --event-qps="50" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311268 4802 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311276 4802 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311284 4802 flags.go:64] FLAG: --eviction-hard="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311293 4802 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311301 4802 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311307 4802 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311314 4802 flags.go:64] FLAG: --eviction-soft="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311322 4802 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311329 4802 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311336 4802 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311343 4802 flags.go:64] FLAG: --experimental-mounter-path="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311351 4802 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311358 4802 flags.go:64] FLAG: --fail-swap-on="true" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311365 4802 flags.go:64] FLAG: --feature-gates="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311374 4802 flags.go:64] FLAG: --file-check-frequency="20s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311382 4802 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311389 4802 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311397 4802 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311404 4802 flags.go:64] FLAG: --healthz-port="10248" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311412 4802 flags.go:64] FLAG: --help="false" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311419 4802 flags.go:64] FLAG: --hostname-override="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311426 4802 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311434 4802 flags.go:64] FLAG: --http-check-frequency="20s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311441 4802 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311448 4802 flags.go:64] FLAG: --image-credential-provider-config="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311457 4802 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311464 4802 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311473 4802 flags.go:64] FLAG: --image-service-endpoint="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311480 4802 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311487 4802 flags.go:64] FLAG: --kube-api-burst="100" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311495 4802 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311502 4802 flags.go:64] FLAG: --kube-api-qps="50" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311509 4802 flags.go:64] FLAG: --kube-reserved="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311539 4802 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311548 4802 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311557 4802 flags.go:64] FLAG: --kubelet-cgroups="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311563 4802 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311571 4802 flags.go:64] FLAG: --lock-file="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311578 4802 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311585 4802 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311593 4802 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311605 4802 flags.go:64] FLAG: --log-json-split-stream="false" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311612 4802 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311620 4802 flags.go:64] FLAG: --log-text-split-stream="false" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311626 4802 flags.go:64] FLAG: --logging-format="text" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311633 4802 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311641 4802 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311648 4802 flags.go:64] FLAG: --manifest-url="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311656 4802 flags.go:64] FLAG: --manifest-url-header="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311666 4802 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311674 4802 flags.go:64] FLAG: --max-open-files="1000000" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311692 4802 flags.go:64] FLAG: --max-pods="110" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311700 4802 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311707 4802 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311714 4802 flags.go:64] FLAG: --memory-manager-policy="None" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311721 4802 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311729 4802 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311738 4802 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311745 4802 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311762 4802 flags.go:64] FLAG: --node-status-max-images="50" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311768 4802 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311776 4802 flags.go:64] FLAG: --oom-score-adj="-999" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311783 4802 flags.go:64] FLAG: --pod-cidr="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311789 4802 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311803 4802 flags.go:64] FLAG: --pod-manifest-path="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311809 4802 flags.go:64] FLAG: --pod-max-pids="-1" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311815 4802 flags.go:64] FLAG: --pods-per-core="0" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311821 4802 flags.go:64] FLAG: --port="10250" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311852 4802 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311866 4802 flags.go:64] FLAG: --provider-id="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311871 4802 flags.go:64] FLAG: --qos-reserved="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311878 4802 flags.go:64] FLAG: --read-only-port="10255" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311884 4802 flags.go:64] FLAG: --register-node="true" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311890 4802 flags.go:64] FLAG: --register-schedulable="true" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311895 4802 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311910 4802 flags.go:64] FLAG: --registry-burst="10" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311916 4802 flags.go:64] FLAG: --registry-qps="5" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311921 4802 flags.go:64] FLAG: --reserved-cpus="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311927 4802 flags.go:64] FLAG: --reserved-memory="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311934 4802 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311940 4802 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311946 4802 flags.go:64] FLAG: --rotate-certificates="false" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311951 4802 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311957 4802 flags.go:64] FLAG: --runonce="false" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311962 4802 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311968 4802 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311974 4802 flags.go:64] FLAG: --seccomp-default="false" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311981 4802 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311987 4802 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.311995 4802 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312003 4802 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312010 4802 flags.go:64] FLAG: --storage-driver-password="root" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312017 4802 flags.go:64] FLAG: --storage-driver-secure="false" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312024 4802 flags.go:64] FLAG: --storage-driver-table="stats" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312031 4802 flags.go:64] FLAG: --storage-driver-user="root" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312038 4802 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312045 4802 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312052 4802 flags.go:64] FLAG: --system-cgroups="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312059 4802 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312071 4802 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312077 4802 flags.go:64] FLAG: --tls-cert-file="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312084 4802 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312091 4802 flags.go:64] FLAG: --tls-min-version="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312096 4802 flags.go:64] FLAG: --tls-private-key-file="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312102 4802 flags.go:64] FLAG: --topology-manager-policy="none" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312118 4802 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312148 4802 flags.go:64] FLAG: --topology-manager-scope="container" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312155 4802 flags.go:64] FLAG: --v="2" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312165 4802 flags.go:64] FLAG: --version="false" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312174 4802 flags.go:64] FLAG: --vmodule="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312185 4802 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.312193 4802 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313191 4802 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313204 4802 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313209 4802 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313214 4802 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313227 4802 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313232 4802 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313237 4802 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313242 4802 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313247 4802 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313252 4802 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313257 4802 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313262 4802 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313267 4802 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313272 4802 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313276 4802 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313281 4802 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313286 4802 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313291 4802 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313296 4802 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313302 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313306 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313311 4802 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313318 4802 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313325 4802 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313330 4802 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313336 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313341 4802 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313347 4802 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313352 4802 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313358 4802 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313364 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313369 4802 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313373 4802 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313379 4802 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313384 4802 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313389 4802 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313393 4802 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313398 4802 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313403 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313408 4802 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313413 4802 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313419 4802 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313426 4802 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313432 4802 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313437 4802 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313443 4802 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313448 4802 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313453 4802 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313459 4802 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313465 4802 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313471 4802 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313482 4802 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313488 4802 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313494 4802 feature_gate.go:330] unrecognized feature gate: Example Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313500 4802 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313506 4802 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313512 4802 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313518 4802 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313526 4802 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313533 4802 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313538 4802 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313543 4802 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313548 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313556 4802 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313564 4802 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313572 4802 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313579 4802 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313585 4802 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313591 4802 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313598 4802 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.313605 4802 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.313623 4802 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.329067 4802 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.329109 4802 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329223 4802 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329234 4802 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329240 4802 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329246 4802 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329250 4802 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329254 4802 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329258 4802 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329263 4802 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329267 4802 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329271 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329275 4802 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329280 4802 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329284 4802 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329288 4802 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329293 4802 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329299 4802 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329304 4802 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329308 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329313 4802 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329318 4802 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329322 4802 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329326 4802 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329330 4802 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329335 4802 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329338 4802 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329342 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329345 4802 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329350 4802 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329354 4802 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329365 4802 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329369 4802 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329374 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329378 4802 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329383 4802 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329388 4802 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329393 4802 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329398 4802 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329403 4802 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329407 4802 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329411 4802 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329416 4802 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329420 4802 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329425 4802 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329429 4802 feature_gate.go:330] unrecognized feature gate: Example Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329433 4802 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329437 4802 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329441 4802 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329444 4802 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329448 4802 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329452 4802 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329455 4802 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329458 4802 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329463 4802 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329467 4802 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329470 4802 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329474 4802 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329478 4802 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329482 4802 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329487 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329490 4802 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329494 4802 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329497 4802 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329501 4802 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329504 4802 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329508 4802 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329511 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329515 4802 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329519 4802 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329522 4802 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329526 4802 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329529 4802 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.329535 4802 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329656 4802 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329664 4802 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329669 4802 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329672 4802 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329676 4802 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329680 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329683 4802 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329687 4802 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329693 4802 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329698 4802 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329702 4802 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329706 4802 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329711 4802 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329716 4802 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329723 4802 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329729 4802 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329734 4802 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329738 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329743 4802 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329747 4802 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329753 4802 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329760 4802 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329765 4802 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329770 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329775 4802 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329779 4802 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329783 4802 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329788 4802 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329792 4802 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329797 4802 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329801 4802 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329805 4802 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329809 4802 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329814 4802 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329821 4802 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329828 4802 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329834 4802 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329840 4802 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329845 4802 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329849 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329853 4802 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329858 4802 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329862 4802 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329867 4802 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329872 4802 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329877 4802 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329881 4802 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329885 4802 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329889 4802 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329894 4802 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329898 4802 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329902 4802 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329906 4802 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329910 4802 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329914 4802 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329919 4802 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329924 4802 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329930 4802 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329934 4802 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329939 4802 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329944 4802 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329948 4802 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329953 4802 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329957 4802 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329961 4802 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329966 4802 feature_gate.go:330] unrecognized feature gate: Example Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329970 4802 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329975 4802 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329981 4802 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329986 4802 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.329991 4802 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.329998 4802 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.331102 4802 server.go:940] "Client rotation is on, will bootstrap in background" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.335031 4802 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.335099 4802 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.336649 4802 server.go:997] "Starting client certificate rotation" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.336672 4802 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.337884 4802 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-03 18:46:13.164375948 +0000 UTC Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.338065 4802 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 937h59m15.826318846s for next certificate rotation Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.365875 4802 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.367826 4802 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.385278 4802 log.go:25] "Validated CRI v1 runtime API" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.415314 4802 log.go:25] "Validated CRI v1 image API" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.417226 4802 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.422525 4802 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-25-16-41-50-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.422549 4802 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.437104 4802 manager.go:217] Machine: {Timestamp:2025-11-25 16:46:57.435489673 +0000 UTC m=+0.579836879 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:7331689e-1a93-42fa-8eca-f0913b387d6b BootID:190f8f70-36fb-445d-a11a-52e80e3afc28 Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:f2:db:c6 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:f2:db:c6 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:1d:9f:9d Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:9c:9d:ae Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:da:a3:12 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:69:0e:14 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ea:0c:69:47:ad:52 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ce:78:d6:3c:3c:9a Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.437476 4802 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.437666 4802 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.438372 4802 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.438565 4802 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.438591 4802 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.438780 4802 topology_manager.go:138] "Creating topology manager with none policy" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.438790 4802 container_manager_linux.go:303] "Creating device plugin manager" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.439278 4802 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.439330 4802 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.439517 4802 state_mem.go:36] "Initialized new in-memory state store" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.439617 4802 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.444214 4802 kubelet.go:418] "Attempting to sync node with API server" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.444241 4802 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.444258 4802 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.444274 4802 kubelet.go:324] "Adding apiserver pod source" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.444289 4802 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.449157 4802 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.450341 4802 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.453056 4802 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.454495 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.454526 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.454536 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.454549 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.454569 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.454581 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.454592 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.454610 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.454620 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.454640 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.454655 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.454665 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.455758 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Nov 25 16:46:57 crc kubenswrapper[4802]: E1125 16:46:57.455828 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.455758 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Nov 25 16:46:57 crc kubenswrapper[4802]: E1125 16:46:57.455881 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.456627 4802 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.457179 4802 server.go:1280] "Started kubelet" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.457402 4802 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.457479 4802 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.457908 4802 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 25 16:46:57 crc systemd[1]: Started Kubernetes Kubelet. Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.458952 4802 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.459944 4802 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.459971 4802 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.460005 4802 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 05:37:17.181157673 +0000 UTC Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.460164 4802 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.460197 4802 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.460261 4802 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 25 16:46:57 crc kubenswrapper[4802]: E1125 16:46:57.460191 4802 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.461413 4802 server.go:460] "Adding debug handlers to kubelet server" Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.462320 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Nov 25 16:46:57 crc kubenswrapper[4802]: E1125 16:46:57.462373 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Nov 25 16:46:57 crc kubenswrapper[4802]: E1125 16:46:57.462322 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" interval="200ms" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.465374 4802 factory.go:55] Registering systemd factory Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.465396 4802 factory.go:221] Registration of the systemd container factory successfully Nov 25 16:46:57 crc kubenswrapper[4802]: E1125 16:46:57.465736 4802 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.245:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b4dcbf6a0ca41 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 16:46:57.457113665 +0000 UTC m=+0.601460871,LastTimestamp:2025-11-25 16:46:57.457113665 +0000 UTC m=+0.601460871,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.466782 4802 factory.go:153] Registering CRI-O factory Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.466793 4802 factory.go:221] Registration of the crio container factory successfully Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.466845 4802 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.466863 4802 factory.go:103] Registering Raw factory Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.466879 4802 manager.go:1196] Started watching for new ooms in manager Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.467359 4802 manager.go:319] Starting recovery of all containers Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.476848 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477178 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477191 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477200 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477210 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477220 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477229 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477238 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477250 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477259 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477270 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477280 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477289 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477302 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477313 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477323 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477332 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477340 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477349 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477359 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477370 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477379 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.477389 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479380 4802 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479403 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479413 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479424 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479436 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479446 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479457 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479466 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479473 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479485 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479494 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479504 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479513 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479539 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479548 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479556 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479565 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479573 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479582 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479591 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479600 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479629 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479642 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479653 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479663 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479673 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479684 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479694 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479706 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479731 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479744 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479756 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479767 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479777 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479788 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479799 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479810 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479819 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479829 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479840 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479849 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479859 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479868 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479877 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479887 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479895 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479905 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479916 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479926 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479935 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479945 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479955 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479973 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479983 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.479993 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480002 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480011 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480020 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480030 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480040 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480050 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480059 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480069 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480080 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480089 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480097 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480105 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480134 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480148 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480160 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480173 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480184 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480193 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480202 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480211 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480220 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480228 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480236 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480247 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480256 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480268 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480276 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480290 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480300 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480309 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480320 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480331 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480340 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480350 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480358 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480368 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480379 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480388 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480399 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480408 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480417 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480443 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480452 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480462 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480471 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480481 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480492 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480502 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480511 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480521 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480531 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480542 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480550 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480559 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480568 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480577 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480587 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480598 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480609 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480634 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480643 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480654 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480663 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480678 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480686 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480695 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480704 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480713 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480722 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480730 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480740 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480749 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480758 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480767 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480777 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480785 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480794 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480803 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480812 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480822 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480831 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480840 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480848 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480857 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480869 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480878 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480886 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480896 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480905 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480918 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480927 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480937 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480946 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480955 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480964 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480975 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480984 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.480992 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481001 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481010 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481017 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481028 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481036 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481045 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481054 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481063 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481072 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481081 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481091 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481101 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481109 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481221 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481232 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481240 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481249 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481258 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481267 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481276 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481285 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481295 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481302 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481313 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481322 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481330 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481339 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481348 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481356 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481366 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481373 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481382 4802 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481391 4802 reconstruct.go:97] "Volume reconstruction finished" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.481397 4802 reconciler.go:26] "Reconciler: start to sync state" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.489425 4802 manager.go:324] Recovery completed Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.498952 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.500862 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.500902 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.500911 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.501416 4802 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.501526 4802 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.501551 4802 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.501581 4802 state_mem.go:36] "Initialized new in-memory state store" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.503248 4802 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.503281 4802 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.503303 4802 kubelet.go:2335] "Starting kubelet main sync loop" Nov 25 16:46:57 crc kubenswrapper[4802]: E1125 16:46:57.503341 4802 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 25 16:46:57 crc kubenswrapper[4802]: W1125 16:46:57.503945 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Nov 25 16:46:57 crc kubenswrapper[4802]: E1125 16:46:57.504026 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.519667 4802 policy_none.go:49] "None policy: Start" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.520709 4802 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.520738 4802 state_mem.go:35] "Initializing new in-memory state store" Nov 25 16:46:57 crc kubenswrapper[4802]: E1125 16:46:57.560470 4802 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.574161 4802 manager.go:334] "Starting Device Plugin manager" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.574206 4802 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.574220 4802 server.go:79] "Starting device plugin registration server" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.574646 4802 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.574666 4802 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.574799 4802 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.574922 4802 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.574930 4802 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 25 16:46:57 crc kubenswrapper[4802]: E1125 16:46:57.583496 4802 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.604335 4802 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.604401 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.605102 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.605153 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.605167 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.605251 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.605490 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.605603 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.605925 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.605968 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.605980 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.606186 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.606313 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.606383 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.607300 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.607322 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.607331 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.609039 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.609060 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.609096 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.609108 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.609070 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.609229 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.609285 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.609408 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.609448 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.610247 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.610284 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.610321 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.610292 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.610415 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.610438 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.610449 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.610579 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.610619 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.611188 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.611207 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.611216 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.611330 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.611348 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.612335 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.612381 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.612393 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.612401 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.612419 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.612428 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:57 crc kubenswrapper[4802]: E1125 16:46:57.665470 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" interval="400ms" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.675667 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.676965 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.677027 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.677044 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.677078 4802 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 16:46:57 crc kubenswrapper[4802]: E1125 16:46:57.677622 4802 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.245:6443: connect: connection refused" node="crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.683835 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.683870 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.683889 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.683907 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.683924 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.683938 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.684051 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.684152 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.684206 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.684255 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.684287 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.684304 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.684319 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.684335 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.684349 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.785303 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.785523 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.785683 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.785791 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.785810 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.785826 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.785840 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.785859 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.785875 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.785916 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.785935 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.785942 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.785948 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.785991 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.785966 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.786009 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.785952 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.785994 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.786096 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.785980 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.786066 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.786006 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.786062 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.786213 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.786255 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.786259 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.786289 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.786305 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.786274 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.786395 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.878375 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.879562 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.879631 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.879641 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.879667 4802 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 16:46:57 crc kubenswrapper[4802]: E1125 16:46:57.880357 4802 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.245:6443: connect: connection refused" node="crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.950374 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.965417 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.972703 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.990111 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 16:46:57 crc kubenswrapper[4802]: I1125 16:46:57.995318 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 16:46:58 crc kubenswrapper[4802]: W1125 16:46:58.003866 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-94ef7d9d31f4319423101408bccd1978e6bb346dd6ec5319c368925771733ec2 WatchSource:0}: Error finding container 94ef7d9d31f4319423101408bccd1978e6bb346dd6ec5319c368925771733ec2: Status 404 returned error can't find the container with id 94ef7d9d31f4319423101408bccd1978e6bb346dd6ec5319c368925771733ec2 Nov 25 16:46:58 crc kubenswrapper[4802]: W1125 16:46:58.004707 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-458c62753889a52a415763d97b3d2ac8d25dbd9f5426c0ab44ec422ffe140f50 WatchSource:0}: Error finding container 458c62753889a52a415763d97b3d2ac8d25dbd9f5426c0ab44ec422ffe140f50: Status 404 returned error can't find the container with id 458c62753889a52a415763d97b3d2ac8d25dbd9f5426c0ab44ec422ffe140f50 Nov 25 16:46:58 crc kubenswrapper[4802]: W1125 16:46:58.010163 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-7e9defc479f22b9601f39b85ca619a21c3c62ec5a11801a6bb9d59ca862c07ef WatchSource:0}: Error finding container 7e9defc479f22b9601f39b85ca619a21c3c62ec5a11801a6bb9d59ca862c07ef: Status 404 returned error can't find the container with id 7e9defc479f22b9601f39b85ca619a21c3c62ec5a11801a6bb9d59ca862c07ef Nov 25 16:46:58 crc kubenswrapper[4802]: W1125 16:46:58.018533 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-01842082803a1a4b14ee7ab4c9ba579fea6a355d0edae0df65a27b44f1a35227 WatchSource:0}: Error finding container 01842082803a1a4b14ee7ab4c9ba579fea6a355d0edae0df65a27b44f1a35227: Status 404 returned error can't find the container with id 01842082803a1a4b14ee7ab4c9ba579fea6a355d0edae0df65a27b44f1a35227 Nov 25 16:46:58 crc kubenswrapper[4802]: W1125 16:46:58.022631 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-0dd71452de4049acccc377b09db7d37e018f7eeb71e23069b23cf09f3aa01964 WatchSource:0}: Error finding container 0dd71452de4049acccc377b09db7d37e018f7eeb71e23069b23cf09f3aa01964: Status 404 returned error can't find the container with id 0dd71452de4049acccc377b09db7d37e018f7eeb71e23069b23cf09f3aa01964 Nov 25 16:46:58 crc kubenswrapper[4802]: E1125 16:46:58.066098 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" interval="800ms" Nov 25 16:46:58 crc kubenswrapper[4802]: I1125 16:46:58.281164 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:58 crc kubenswrapper[4802]: I1125 16:46:58.282845 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:58 crc kubenswrapper[4802]: I1125 16:46:58.282889 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:58 crc kubenswrapper[4802]: I1125 16:46:58.282899 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:58 crc kubenswrapper[4802]: I1125 16:46:58.282925 4802 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 16:46:58 crc kubenswrapper[4802]: E1125 16:46:58.283382 4802 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.245:6443: connect: connection refused" node="crc" Nov 25 16:46:58 crc kubenswrapper[4802]: W1125 16:46:58.383386 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Nov 25 16:46:58 crc kubenswrapper[4802]: E1125 16:46:58.383544 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Nov 25 16:46:58 crc kubenswrapper[4802]: I1125 16:46:58.460130 4802 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 05:53:25.938603576 +0000 UTC Nov 25 16:46:58 crc kubenswrapper[4802]: I1125 16:46:58.460215 4802 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1021h6m27.478391727s for next certificate rotation Nov 25 16:46:58 crc kubenswrapper[4802]: I1125 16:46:58.460672 4802 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Nov 25 16:46:58 crc kubenswrapper[4802]: W1125 16:46:58.491137 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Nov 25 16:46:58 crc kubenswrapper[4802]: E1125 16:46:58.491205 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Nov 25 16:46:58 crc kubenswrapper[4802]: W1125 16:46:58.503465 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Nov 25 16:46:58 crc kubenswrapper[4802]: E1125 16:46:58.503550 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Nov 25 16:46:58 crc kubenswrapper[4802]: I1125 16:46:58.509956 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0dd71452de4049acccc377b09db7d37e018f7eeb71e23069b23cf09f3aa01964"} Nov 25 16:46:58 crc kubenswrapper[4802]: I1125 16:46:58.511218 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"01842082803a1a4b14ee7ab4c9ba579fea6a355d0edae0df65a27b44f1a35227"} Nov 25 16:46:58 crc kubenswrapper[4802]: I1125 16:46:58.512782 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7e9defc479f22b9601f39b85ca619a21c3c62ec5a11801a6bb9d59ca862c07ef"} Nov 25 16:46:58 crc kubenswrapper[4802]: I1125 16:46:58.513437 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"458c62753889a52a415763d97b3d2ac8d25dbd9f5426c0ab44ec422ffe140f50"} Nov 25 16:46:58 crc kubenswrapper[4802]: I1125 16:46:58.514012 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"94ef7d9d31f4319423101408bccd1978e6bb346dd6ec5319c368925771733ec2"} Nov 25 16:46:58 crc kubenswrapper[4802]: W1125 16:46:58.762692 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Nov 25 16:46:58 crc kubenswrapper[4802]: E1125 16:46:58.762773 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Nov 25 16:46:58 crc kubenswrapper[4802]: E1125 16:46:58.867637 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" interval="1.6s" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.083523 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.084363 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.084399 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.084410 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.084430 4802 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 16:46:59 crc kubenswrapper[4802]: E1125 16:46:59.084864 4802 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.245:6443: connect: connection refused" node="crc" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.459803 4802 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.519516 4802 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45" exitCode=0 Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.519585 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45"} Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.519627 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.520936 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.520971 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.520985 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.524455 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f"} Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.524490 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e"} Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.524503 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e"} Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.524515 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89"} Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.524559 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.525753 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.525776 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.525787 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.526269 4802 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375" exitCode=0 Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.526304 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.526362 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375"} Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.527440 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.527472 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.527483 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.527697 4802 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d1185051e966f5abc81230ae57bcbdf09f308bf184503734277d3b593c9a9962" exitCode=0 Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.527761 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d1185051e966f5abc81230ae57bcbdf09f308bf184503734277d3b593c9a9962"} Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.527899 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.528608 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.528870 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.528913 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.528924 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.530309 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.530342 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.530355 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.530767 4802 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="74e0f0d4a5a5b635060b11f8b506df6000fe73f589125d64bdc0450dec47ccbe" exitCode=0 Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.530824 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.530847 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"74e0f0d4a5a5b635060b11f8b506df6000fe73f589125d64bdc0450dec47ccbe"} Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.531613 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.531641 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:46:59 crc kubenswrapper[4802]: I1125 16:46:59.531653 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:00 crc kubenswrapper[4802]: W1125 16:47:00.062976 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Nov 25 16:47:00 crc kubenswrapper[4802]: E1125 16:47:00.063052 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Nov 25 16:47:00 crc kubenswrapper[4802]: W1125 16:47:00.270197 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Nov 25 16:47:00 crc kubenswrapper[4802]: E1125 16:47:00.270291 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.460196 4802 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Nov 25 16:47:00 crc kubenswrapper[4802]: E1125 16:47:00.468966 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" interval="3.2s" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.534701 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"db167a028aae7e1be86605fcfb5a716b35c6e34d600b9675fb2c26b88bc52eb3"} Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.534758 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.535942 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.535978 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.535988 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.538081 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c4bee24ceffdebad39410dfdca5f8a83c8664f6b4c18cc2495f47f47162d116a"} Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.538110 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ceb75f902fead2f9945afbecfdfd7b7fd35c5eaa4a05059a2038fb08e07875ee"} Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.538136 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"060bdec5b0ff95b1ac4dcdb60f4bbb811ddeff0c93e74e373640bf65cb1d5ec8"} Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.538169 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.539141 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.539174 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.539189 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.541700 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36"} Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.541725 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094"} Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.541735 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80"} Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.541744 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488"} Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.541753 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875"} Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.541758 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.542430 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.542452 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.542460 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.545453 4802 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="47bb2296db0230f43e93667552e76ad7345e9328d82a5424b4ae5ac6800aac09" exitCode=0 Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.545476 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"47bb2296db0230f43e93667552e76ad7345e9328d82a5424b4ae5ac6800aac09"} Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.545539 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.545557 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.546235 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.546270 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.546285 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.547033 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.547068 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.547083 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.667009 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.685039 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.686260 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.686324 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.686335 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:00 crc kubenswrapper[4802]: I1125 16:47:00.686384 4802 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 16:47:00 crc kubenswrapper[4802]: E1125 16:47:00.686990 4802 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.245:6443: connect: connection refused" node="crc" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.348582 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.549945 4802 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e7fc75aef3030e6cdf85e714435a8eacf3c017245fb2303b20d53f8a0ec4ab07" exitCode=0 Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.550000 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e7fc75aef3030e6cdf85e714435a8eacf3c017245fb2303b20d53f8a0ec4ab07"} Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.550042 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.550050 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.550076 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.550141 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.550077 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.550588 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.550078 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.550880 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.550899 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.550907 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.550957 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.550971 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.550978 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.551080 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.551098 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.551108 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.551702 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.551723 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.551730 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.551743 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.551756 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.551763 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:01 crc kubenswrapper[4802]: I1125 16:47:01.705017 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 16:47:02 crc kubenswrapper[4802]: I1125 16:47:02.558706 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ec9e33dc5d83ad1c5bfd900ff3ad2bced791b5440a71a1476b5a63ff6c26c9dc"} Nov 25 16:47:02 crc kubenswrapper[4802]: I1125 16:47:02.558777 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9724a6b84b44bc563f26f7a749bd637590ae7ab59730a02bc305f743025a27c5"} Nov 25 16:47:02 crc kubenswrapper[4802]: I1125 16:47:02.558798 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9c18705be2c15f0b702df411520620821d104db62e782ea41095a7c10c027eaf"} Nov 25 16:47:02 crc kubenswrapper[4802]: I1125 16:47:02.558813 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8b4c6a68d8d23da5cb447e9089c119cd555a008d4a3672a8ad427c190b6c0c06"} Nov 25 16:47:02 crc kubenswrapper[4802]: I1125 16:47:02.558829 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:02 crc kubenswrapper[4802]: I1125 16:47:02.558919 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:02 crc kubenswrapper[4802]: I1125 16:47:02.558829 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3e0756affdcfa3981646ba34881bec0c0721d10e1fe43cc7978a9b625e9fef2f"} Nov 25 16:47:02 crc kubenswrapper[4802]: I1125 16:47:02.558826 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 16:47:02 crc kubenswrapper[4802]: I1125 16:47:02.559042 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:02 crc kubenswrapper[4802]: I1125 16:47:02.560308 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:02 crc kubenswrapper[4802]: I1125 16:47:02.560338 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:02 crc kubenswrapper[4802]: I1125 16:47:02.560342 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:02 crc kubenswrapper[4802]: I1125 16:47:02.560371 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:02 crc kubenswrapper[4802]: I1125 16:47:02.560385 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:02 crc kubenswrapper[4802]: I1125 16:47:02.560396 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:02 crc kubenswrapper[4802]: I1125 16:47:02.560408 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:02 crc kubenswrapper[4802]: I1125 16:47:02.560350 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:02 crc kubenswrapper[4802]: I1125 16:47:02.560417 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.146654 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.147841 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.153845 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.270191 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.349877 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.560648 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.560678 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.560706 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.560775 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.562215 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.562247 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.562213 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.562259 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.562264 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.562376 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.562388 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.562280 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.562415 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.887329 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.888553 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.888582 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.888591 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:03 crc kubenswrapper[4802]: I1125 16:47:03.888611 4802 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 16:47:04 crc kubenswrapper[4802]: I1125 16:47:04.562389 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:04 crc kubenswrapper[4802]: I1125 16:47:04.562407 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:04 crc kubenswrapper[4802]: I1125 16:47:04.563407 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:04 crc kubenswrapper[4802]: I1125 16:47:04.563448 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:04 crc kubenswrapper[4802]: I1125 16:47:04.563463 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:04 crc kubenswrapper[4802]: I1125 16:47:04.563472 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:04 crc kubenswrapper[4802]: I1125 16:47:04.563480 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:04 crc kubenswrapper[4802]: I1125 16:47:04.563490 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:05 crc kubenswrapper[4802]: I1125 16:47:05.785364 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 16:47:05 crc kubenswrapper[4802]: I1125 16:47:05.785808 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:05 crc kubenswrapper[4802]: I1125 16:47:05.786895 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:05 crc kubenswrapper[4802]: I1125 16:47:05.786926 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:05 crc kubenswrapper[4802]: I1125 16:47:05.786941 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:05 crc kubenswrapper[4802]: I1125 16:47:05.794459 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 16:47:05 crc kubenswrapper[4802]: I1125 16:47:05.795663 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:05 crc kubenswrapper[4802]: I1125 16:47:05.796829 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:05 crc kubenswrapper[4802]: I1125 16:47:05.796874 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:05 crc kubenswrapper[4802]: I1125 16:47:05.796888 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:05 crc kubenswrapper[4802]: I1125 16:47:05.799162 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 25 16:47:05 crc kubenswrapper[4802]: I1125 16:47:05.799316 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:05 crc kubenswrapper[4802]: I1125 16:47:05.800409 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:05 crc kubenswrapper[4802]: I1125 16:47:05.800482 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:05 crc kubenswrapper[4802]: I1125 16:47:05.800499 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:06 crc kubenswrapper[4802]: I1125 16:47:06.270902 4802 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 16:47:06 crc kubenswrapper[4802]: I1125 16:47:06.270981 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 16:47:07 crc kubenswrapper[4802]: E1125 16:47:07.584530 4802 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 16:47:11 crc kubenswrapper[4802]: W1125 16:47:11.268783 4802 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 25 16:47:11 crc kubenswrapper[4802]: I1125 16:47:11.268880 4802 trace.go:236] Trace[1394063924]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 16:47:01.267) (total time: 10001ms): Nov 25 16:47:11 crc kubenswrapper[4802]: Trace[1394063924]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (16:47:11.268) Nov 25 16:47:11 crc kubenswrapper[4802]: Trace[1394063924]: [10.001138469s] [10.001138469s] END Nov 25 16:47:11 crc kubenswrapper[4802]: E1125 16:47:11.268901 4802 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 25 16:47:11 crc kubenswrapper[4802]: I1125 16:47:11.349384 4802 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": context deadline exceeded" start-of-body= Nov 25 16:47:11 crc kubenswrapper[4802]: I1125 16:47:11.349451 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": context deadline exceeded" Nov 25 16:47:11 crc kubenswrapper[4802]: I1125 16:47:11.461386 4802 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 25 16:47:11 crc kubenswrapper[4802]: I1125 16:47:11.628440 4802 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 25 16:47:11 crc kubenswrapper[4802]: I1125 16:47:11.628518 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 16:47:11 crc kubenswrapper[4802]: I1125 16:47:11.710830 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 16:47:11 crc kubenswrapper[4802]: I1125 16:47:11.710957 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:11 crc kubenswrapper[4802]: I1125 16:47:11.711952 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:11 crc kubenswrapper[4802]: I1125 16:47:11.711991 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:11 crc kubenswrapper[4802]: I1125 16:47:11.712007 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:13 crc kubenswrapper[4802]: I1125 16:47:13.394031 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 25 16:47:13 crc kubenswrapper[4802]: I1125 16:47:13.394308 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:13 crc kubenswrapper[4802]: I1125 16:47:13.398616 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:13 crc kubenswrapper[4802]: I1125 16:47:13.398661 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:13 crc kubenswrapper[4802]: I1125 16:47:13.398672 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:13 crc kubenswrapper[4802]: I1125 16:47:13.415948 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 25 16:47:13 crc kubenswrapper[4802]: I1125 16:47:13.582452 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:13 crc kubenswrapper[4802]: I1125 16:47:13.583591 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:13 crc kubenswrapper[4802]: I1125 16:47:13.583662 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:13 crc kubenswrapper[4802]: I1125 16:47:13.583687 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:15 crc kubenswrapper[4802]: I1125 16:47:15.739481 4802 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.271079 4802 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.271261 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.355379 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.364349 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.455042 4802 apiserver.go:52] "Watching apiserver" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.461038 4802 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.461516 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.462049 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.462048 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.462542 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.462573 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.462556 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.462647 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.462672 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.462681 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.463001 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.465184 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.466420 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.466605 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.467169 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.467228 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.467295 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.467346 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.467310 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.468275 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.495650 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.505202 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.515021 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.521912 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.530698 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.538178 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.545392 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.553243 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.561659 4802 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.588600 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.594370 4802 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.619420 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.619557 4802 trace.go:236] Trace[2058211478]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 16:47:05.257) (total time: 11361ms): Nov 25 16:47:16 crc kubenswrapper[4802]: Trace[2058211478]: ---"Objects listed" error: 11361ms (16:47:16.619) Nov 25 16:47:16 crc kubenswrapper[4802]: Trace[2058211478]: [11.361594059s] [11.361594059s] END Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.619574 4802 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.621701 4802 trace.go:236] Trace[1398550927]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 16:47:04.366) (total time: 12255ms): Nov 25 16:47:16 crc kubenswrapper[4802]: Trace[1398550927]: ---"Objects listed" error: 12255ms (16:47:16.621) Nov 25 16:47:16 crc kubenswrapper[4802]: Trace[1398550927]: [12.255325156s] [12.255325156s] END Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.621724 4802 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.622272 4802 trace.go:236] Trace[1115085885]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 16:47:01.659) (total time: 14962ms): Nov 25 16:47:16 crc kubenswrapper[4802]: Trace[1115085885]: ---"Objects listed" error: 14962ms (16:47:16.622) Nov 25 16:47:16 crc kubenswrapper[4802]: Trace[1115085885]: [14.962941707s] [14.962941707s] END Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.622284 4802 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.622511 4802 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.622608 4802 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.665087 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.683040 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.694479 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.705177 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.721204 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.722830 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.722890 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.722919 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.722947 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.722972 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.722997 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723023 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723051 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723075 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723102 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723144 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723168 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723192 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723272 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723298 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723324 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723346 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723377 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723403 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723426 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723482 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723507 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723338 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723561 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723586 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723608 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723633 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723655 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723677 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723698 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723719 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723739 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723763 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723790 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723811 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723833 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723860 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723886 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723947 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723970 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724003 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724026 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724047 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724069 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724091 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724112 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724151 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724177 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724202 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724226 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724294 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724408 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724435 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724460 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723585 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724484 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723356 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724510 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724539 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724574 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724598 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724621 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724644 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724669 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724695 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724717 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724745 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724777 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724811 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724842 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724869 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724894 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724917 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724941 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724968 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724993 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725017 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725038 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725064 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725089 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725112 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725160 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725185 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725208 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725233 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725258 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725281 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725304 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725326 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725348 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725371 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725394 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725417 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725441 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725465 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725497 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725521 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725544 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725568 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725591 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725616 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725641 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725664 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725689 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725711 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725735 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728183 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728294 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728385 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728450 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728521 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728570 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728619 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728669 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728717 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728768 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728818 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730221 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730306 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730347 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730371 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730393 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730419 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730443 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730465 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730486 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730507 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730527 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730546 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730566 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730587 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730605 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730627 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730682 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730700 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730720 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730739 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730760 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730778 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730798 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730942 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730968 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731002 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731019 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731039 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731064 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731091 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731115 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731161 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731189 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731215 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731245 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731275 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731303 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731331 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731402 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731430 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731456 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731486 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731517 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731542 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731572 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731606 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731639 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731666 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731713 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731747 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731775 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723522 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723523 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723742 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723806 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723875 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723845 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723960 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723947 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.723981 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724009 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724137 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724249 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724279 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724290 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724461 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724458 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724576 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724581 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724588 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724663 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724759 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724755 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724805 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.732413 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:47:17.232386281 +0000 UTC m=+20.376733477 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.724891 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725058 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725208 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725230 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725532 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.725614 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.727583 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.727667 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.727720 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.727744 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728089 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728103 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728242 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728439 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728429 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728449 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728659 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728769 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.728882 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.729014 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.729455 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.729529 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.729780 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.729876 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730105 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730413 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.730606 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731033 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731298 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731435 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731483 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731695 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.731929 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.732007 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.732629 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.732811 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.732873 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.732921 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.733198 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.733222 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.733282 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.734300 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.735163 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.735405 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.735625 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.735269 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.735868 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736096 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736176 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736218 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736251 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736253 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736282 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736318 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736360 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736389 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736419 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736448 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736480 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736506 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736536 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736569 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736597 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736626 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736654 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736693 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736883 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.738810 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.738943 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.739059 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.739166 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.739251 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.739327 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.739394 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.739462 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.739541 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.739612 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.739781 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.739852 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.739876 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.739900 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.739925 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.739946 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.739979 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.740016 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.740041 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.740062 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.740079 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.740107 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.740148 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.740170 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.740191 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.740280 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.740297 4802 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.740309 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.740320 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.740332 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.740346 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.740356 4802 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.740366 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.740379 4802 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.741250 4802 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.741319 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.741376 4802 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.741435 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.741492 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.741558 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.741619 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.741683 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.741741 4802 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.741797 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.741860 4802 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.741923 4802 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.741982 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.742042 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.742103 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.742207 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.742278 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.742335 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.742925 4802 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743002 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743066 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743196 4802 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743283 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743361 4802 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743428 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743485 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743549 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743613 4802 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743676 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743741 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743802 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743858 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743918 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.744602 4802 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.744634 4802 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.744652 4802 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.744669 4802 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.744681 4802 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.744694 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.744709 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.744722 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736314 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736704 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.736897 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.739308 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.741430 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.741511 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.741801 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.741850 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.742264 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.742289 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.742596 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743187 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743225 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743426 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743523 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743642 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743863 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.743873 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.744050 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.744281 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.744545 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.744821 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.744967 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.744968 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.744990 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.745078 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.745110 4802 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.745649 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.745134 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.745283 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.744552 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.745357 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.745408 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.745904 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.745441 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.745745 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.746539 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.746689 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.746761 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.746904 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.746937 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.747172 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.747264 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.747718 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.747755 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.748479 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.745116 4802 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.751023 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.751553 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.751632 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.751826 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.751909 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.751926 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.752488 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.752764 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.753513 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.754172 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.754523 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.754841 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.755301 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.755565 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.755745 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.755912 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.756501 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.756921 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.757170 4802 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.757241 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:17.257218295 +0000 UTC m=+20.401565571 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.757557 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.758775 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.760626 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.760674 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.760937 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.761539 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.761888 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.762031 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.762063 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.762289 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.762469 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.762658 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.762878 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.763009 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.763098 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.763191 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.763172 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.763227 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.763370 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.763408 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.763429 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.763441 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.763487 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.763652 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.763667 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.763807 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.764409 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.764442 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.764708 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.764798 4802 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.764846 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.764861 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:17.264840745 +0000 UTC m=+20.409187931 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.764934 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.764991 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.765837 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.766278 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.766307 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.766322 4802 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.766378 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:17.266359942 +0000 UTC m=+20.410707248 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766495 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.749413 4802 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766543 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766712 4802 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766737 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766754 4802 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766767 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766780 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766791 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766801 4802 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766812 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766823 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766834 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766847 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766859 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766869 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766878 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766887 4802 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766896 4802 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766906 4802 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766915 4802 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766924 4802 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766932 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766942 4802 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.766953 4802 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.768936 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.769553 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.770511 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.774004 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.776781 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.777904 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.777936 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.777951 4802 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.777995 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: E1125 16:47:16.778231 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:17.27798919 +0000 UTC m=+20.422336516 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.777894 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.778381 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.778835 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.779115 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.779247 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.780272 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.780349 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.780824 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.781254 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.781349 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.781552 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.781584 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.781721 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.781950 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.782059 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.783808 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.783835 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.784572 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.787580 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.790904 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.791155 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.792761 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.795747 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.797593 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.799095 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.810359 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.814007 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.818718 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867695 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867764 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867806 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867817 4802 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867826 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867834 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867842 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867851 4802 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867859 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867866 4802 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867875 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867883 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867891 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867898 4802 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867908 4802 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867920 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867933 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867944 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867954 4802 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867965 4802 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867976 4802 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867987 4802 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.867998 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868010 4802 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868022 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868035 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868047 4802 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868058 4802 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868070 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868082 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868094 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868107 4802 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868137 4802 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868147 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868156 4802 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868164 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868172 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868180 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868188 4802 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868196 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868204 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868212 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868220 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868231 4802 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868239 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868247 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868255 4802 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868263 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868271 4802 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868283 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868293 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868303 4802 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868314 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868324 4802 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868335 4802 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868346 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868359 4802 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868370 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868382 4802 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868393 4802 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868404 4802 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868414 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868425 4802 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868437 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868449 4802 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868460 4802 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868471 4802 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868482 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868493 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868506 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868516 4802 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868524 4802 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868531 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868539 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868548 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868555 4802 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868563 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868572 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868581 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868589 4802 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868598 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868607 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868615 4802 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868622 4802 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868630 4802 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868639 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868648 4802 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868655 4802 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868663 4802 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868670 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868681 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868692 4802 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868702 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868713 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868724 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868735 4802 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868744 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868752 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868760 4802 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868768 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868776 4802 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868785 4802 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868794 4802 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868805 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868816 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868828 4802 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868840 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868851 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868862 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868873 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868884 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868906 4802 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868914 4802 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868922 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868930 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868939 4802 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868947 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868956 4802 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868964 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868972 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868980 4802 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.868988 4802 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.869033 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 16:47:16 crc kubenswrapper[4802]: I1125 16:47:16.869166 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.077869 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.087660 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 16:47:17 crc kubenswrapper[4802]: W1125 16:47:17.088998 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-e6d8a4024ed7b44401b306788ac344398aa877ef5b360f58deead40813755f2d WatchSource:0}: Error finding container e6d8a4024ed7b44401b306788ac344398aa877ef5b360f58deead40813755f2d: Status 404 returned error can't find the container with id e6d8a4024ed7b44401b306788ac344398aa877ef5b360f58deead40813755f2d Nov 25 16:47:17 crc kubenswrapper[4802]: W1125 16:47:17.103905 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-22a34de7a40fc76433e677b2545bf74e59f85fd008a4776ee48326bffceb0741 WatchSource:0}: Error finding container 22a34de7a40fc76433e677b2545bf74e59f85fd008a4776ee48326bffceb0741: Status 404 returned error can't find the container with id 22a34de7a40fc76433e677b2545bf74e59f85fd008a4776ee48326bffceb0741 Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.271159 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.271250 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.271284 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:17 crc kubenswrapper[4802]: E1125 16:47:17.271310 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:47:18.271291091 +0000 UTC m=+21.415638287 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.271347 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:17 crc kubenswrapper[4802]: E1125 16:47:17.271400 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 16:47:17 crc kubenswrapper[4802]: E1125 16:47:17.271406 4802 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 16:47:17 crc kubenswrapper[4802]: E1125 16:47:17.271421 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 16:47:17 crc kubenswrapper[4802]: E1125 16:47:17.271432 4802 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:17 crc kubenswrapper[4802]: E1125 16:47:17.271456 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:18.271446425 +0000 UTC m=+21.415793611 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 16:47:17 crc kubenswrapper[4802]: E1125 16:47:17.271474 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:18.271464425 +0000 UTC m=+21.415811611 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:17 crc kubenswrapper[4802]: E1125 16:47:17.271476 4802 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 16:47:17 crc kubenswrapper[4802]: E1125 16:47:17.271567 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:18.271548587 +0000 UTC m=+21.415895773 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.372023 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:17 crc kubenswrapper[4802]: E1125 16:47:17.372164 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 16:47:17 crc kubenswrapper[4802]: E1125 16:47:17.372189 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 16:47:17 crc kubenswrapper[4802]: E1125 16:47:17.372201 4802 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:17 crc kubenswrapper[4802]: E1125 16:47:17.372256 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:18.372240222 +0000 UTC m=+21.516587408 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.503706 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:17 crc kubenswrapper[4802]: E1125 16:47:17.503829 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.507712 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.508309 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.509047 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.509641 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.510198 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.510665 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.511218 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.511693 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.512267 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.512743 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.513238 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.513984 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.514580 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.515203 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.515681 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.516170 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.516693 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.517088 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.518685 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.519492 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.519532 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.520013 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.520615 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.521143 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.521841 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.522358 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.523000 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.523758 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.524351 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.525038 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.525613 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.526218 4802 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.526340 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.528723 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.529205 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.529591 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.531031 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.531793 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.532387 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.533115 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.533410 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.533834 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.534303 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.534850 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.536666 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.537349 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.538197 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.538719 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.539620 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.540409 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.541316 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.541739 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.542696 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.543434 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.544114 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.544343 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.544665 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.555814 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.566958 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.580774 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.590837 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"22a34de7a40fc76433e677b2545bf74e59f85fd008a4776ee48326bffceb0741"} Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.592099 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611"} Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.592169 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e6d8a4024ed7b44401b306788ac344398aa877ef5b360f58deead40813755f2d"} Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.593592 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8"} Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.593638 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d"} Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.593657 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d87814aac07f788f876cd2b9956c2b1beb5e59c2e4713b635a60b3bb1abcda9d"} Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.598405 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.608267 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.623114 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.635502 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.646050 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.656799 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.668498 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:17 crc kubenswrapper[4802]: I1125 16:47:17.679969 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.281393 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:47:18 crc kubenswrapper[4802]: E1125 16:47:18.281567 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:47:20.281541337 +0000 UTC m=+23.425888523 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.281755 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.281794 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.281813 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:18 crc kubenswrapper[4802]: E1125 16:47:18.281820 4802 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 16:47:18 crc kubenswrapper[4802]: E1125 16:47:18.281955 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 16:47:18 crc kubenswrapper[4802]: E1125 16:47:18.282189 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:20.282156642 +0000 UTC m=+23.426503828 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 16:47:18 crc kubenswrapper[4802]: E1125 16:47:18.282189 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 16:47:18 crc kubenswrapper[4802]: E1125 16:47:18.282228 4802 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:18 crc kubenswrapper[4802]: E1125 16:47:18.282290 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:20.282282366 +0000 UTC m=+23.426629552 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:18 crc kubenswrapper[4802]: E1125 16:47:18.282367 4802 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 16:47:18 crc kubenswrapper[4802]: E1125 16:47:18.282409 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:20.282400179 +0000 UTC m=+23.426747365 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.382724 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:18 crc kubenswrapper[4802]: E1125 16:47:18.382895 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 16:47:18 crc kubenswrapper[4802]: E1125 16:47:18.382929 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 16:47:18 crc kubenswrapper[4802]: E1125 16:47:18.382940 4802 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:18 crc kubenswrapper[4802]: E1125 16:47:18.382997 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:20.38298067 +0000 UTC m=+23.527327846 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.504038 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.504101 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:18 crc kubenswrapper[4802]: E1125 16:47:18.504182 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:18 crc kubenswrapper[4802]: E1125 16:47:18.504251 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.685537 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-dqg6t"] Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.685825 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-dqg6t" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.695703 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-hmbvn"] Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.696028 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.696409 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.696777 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.696794 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.701406 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-qclgb"] Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.701792 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.702008 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.702095 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-h29wc"] Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.702174 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.702428 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.702681 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.702884 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.703634 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.705289 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.705819 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.705914 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.706031 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.706080 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.706338 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.707218 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.714929 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.728237 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.744932 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.757285 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.771387 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.783500 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.786286 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7677a2dc-6b77-4fbc-90e3-9548a2298016-hosts-file\") pod \"node-resolver-dqg6t\" (UID: \"7677a2dc-6b77-4fbc-90e3-9548a2298016\") " pod="openshift-dns/node-resolver-dqg6t" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.786316 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4c3e23e7-2aae-4a81-b64f-4ed182a43d5e-proxy-tls\") pod \"machine-config-daemon-h29wc\" (UID: \"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\") " pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.786333 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-system-cni-dir\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.786346 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-etc-kubernetes\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.786372 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md8dh\" (UniqueName: \"kubernetes.io/projected/97e822d6-58fe-41f4-b08b-3c9b42273307-kube-api-access-md8dh\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.786389 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ghr7\" (UniqueName: \"kubernetes.io/projected/7677a2dc-6b77-4fbc-90e3-9548a2298016-kube-api-access-5ghr7\") pod \"node-resolver-dqg6t\" (UID: \"7677a2dc-6b77-4fbc-90e3-9548a2298016\") " pod="openshift-dns/node-resolver-dqg6t" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.786405 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-multus-cni-dir\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.786525 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.786583 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xc62\" (UniqueName: \"kubernetes.io/projected/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-kube-api-access-4xc62\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.786632 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-os-release\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.786686 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-multus-socket-dir-parent\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.786725 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-host-run-netns\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.786764 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-multus-conf-dir\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.786823 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4c3e23e7-2aae-4a81-b64f-4ed182a43d5e-mcd-auth-proxy-config\") pod \"machine-config-daemon-h29wc\" (UID: \"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\") " pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.786849 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/97e822d6-58fe-41f4-b08b-3c9b42273307-multus-daemon-config\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.786890 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.786948 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-host-var-lib-kubelet\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.786976 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-host-run-multus-certs\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.787000 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-host-var-lib-cni-multus\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.787037 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-cni-binary-copy\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.787063 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-host-var-lib-cni-bin\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.787088 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrbk6\" (UniqueName: \"kubernetes.io/projected/4c3e23e7-2aae-4a81-b64f-4ed182a43d5e-kube-api-access-lrbk6\") pod \"machine-config-daemon-h29wc\" (UID: \"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\") " pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.787110 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-cnibin\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.787147 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-host-run-k8s-cni-cncf-io\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.787179 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-system-cni-dir\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.787205 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-cnibin\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.787227 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-os-release\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.787251 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4c3e23e7-2aae-4a81-b64f-4ed182a43d5e-rootfs\") pod \"machine-config-daemon-h29wc\" (UID: \"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\") " pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.787270 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/97e822d6-58fe-41f4-b08b-3c9b42273307-cni-binary-copy\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.787293 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-hostroot\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.799990 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.811671 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.826650 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.837568 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.850905 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.862446 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.876192 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888358 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-host-var-lib-cni-bin\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888394 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrbk6\" (UniqueName: \"kubernetes.io/projected/4c3e23e7-2aae-4a81-b64f-4ed182a43d5e-kube-api-access-lrbk6\") pod \"machine-config-daemon-h29wc\" (UID: \"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\") " pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888417 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-cnibin\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888441 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-host-run-k8s-cni-cncf-io\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888464 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-system-cni-dir\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888484 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-cnibin\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888511 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-os-release\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888531 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4c3e23e7-2aae-4a81-b64f-4ed182a43d5e-rootfs\") pod \"machine-config-daemon-h29wc\" (UID: \"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\") " pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888552 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/97e822d6-58fe-41f4-b08b-3c9b42273307-cni-binary-copy\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888564 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-cnibin\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888610 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4c3e23e7-2aae-4a81-b64f-4ed182a43d5e-rootfs\") pod \"machine-config-daemon-h29wc\" (UID: \"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\") " pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888632 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-host-run-k8s-cni-cncf-io\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888573 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-cnibin\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888518 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-host-var-lib-cni-bin\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888615 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-hostroot\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888574 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-hostroot\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888745 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7677a2dc-6b77-4fbc-90e3-9548a2298016-hosts-file\") pod \"node-resolver-dqg6t\" (UID: \"7677a2dc-6b77-4fbc-90e3-9548a2298016\") " pod="openshift-dns/node-resolver-dqg6t" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888573 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-system-cni-dir\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888762 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4c3e23e7-2aae-4a81-b64f-4ed182a43d5e-proxy-tls\") pod \"machine-config-daemon-h29wc\" (UID: \"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\") " pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888778 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-system-cni-dir\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888793 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-etc-kubernetes\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888814 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md8dh\" (UniqueName: \"kubernetes.io/projected/97e822d6-58fe-41f4-b08b-3c9b42273307-kube-api-access-md8dh\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888833 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-system-cni-dir\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888833 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ghr7\" (UniqueName: \"kubernetes.io/projected/7677a2dc-6b77-4fbc-90e3-9548a2298016-kube-api-access-5ghr7\") pod \"node-resolver-dqg6t\" (UID: \"7677a2dc-6b77-4fbc-90e3-9548a2298016\") " pod="openshift-dns/node-resolver-dqg6t" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888865 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-etc-kubernetes\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888903 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-multus-cni-dir\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888910 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7677a2dc-6b77-4fbc-90e3-9548a2298016-hosts-file\") pod \"node-resolver-dqg6t\" (UID: \"7677a2dc-6b77-4fbc-90e3-9548a2298016\") " pod="openshift-dns/node-resolver-dqg6t" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888920 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888984 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-multus-cni-dir\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888925 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-os-release\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.888992 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xc62\" (UniqueName: \"kubernetes.io/projected/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-kube-api-access-4xc62\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.889533 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-multus-socket-dir-parent\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.889602 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-multus-socket-dir-parent\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.889704 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.889722 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-host-run-netns\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.889774 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-multus-conf-dir\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.889811 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/97e822d6-58fe-41f4-b08b-3c9b42273307-cni-binary-copy\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.889825 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-host-run-netns\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.889826 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-os-release\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.889869 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4c3e23e7-2aae-4a81-b64f-4ed182a43d5e-mcd-auth-proxy-config\") pod \"machine-config-daemon-h29wc\" (UID: \"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\") " pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.889892 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-multus-conf-dir\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.889914 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.889950 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-os-release\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.889954 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/97e822d6-58fe-41f4-b08b-3c9b42273307-multus-daemon-config\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.889979 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-host-var-lib-kubelet\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.890010 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-host-run-multus-certs\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.890042 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-host-var-lib-cni-multus\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.890076 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-cni-binary-copy\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.890642 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.890637 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/97e822d6-58fe-41f4-b08b-3c9b42273307-multus-daemon-config\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.890682 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-host-run-multus-certs\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.890715 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-host-var-lib-cni-multus\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.890754 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/97e822d6-58fe-41f4-b08b-3c9b42273307-host-var-lib-kubelet\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.890770 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4c3e23e7-2aae-4a81-b64f-4ed182a43d5e-mcd-auth-proxy-config\") pod \"machine-config-daemon-h29wc\" (UID: \"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\") " pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.890835 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-cni-binary-copy\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.897104 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4c3e23e7-2aae-4a81-b64f-4ed182a43d5e-proxy-tls\") pod \"machine-config-daemon-h29wc\" (UID: \"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\") " pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.897577 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.909417 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrbk6\" (UniqueName: \"kubernetes.io/projected/4c3e23e7-2aae-4a81-b64f-4ed182a43d5e-kube-api-access-lrbk6\") pod \"machine-config-daemon-h29wc\" (UID: \"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\") " pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.910702 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xc62\" (UniqueName: \"kubernetes.io/projected/32628afd-a82e-48a8-a3d2-cfb23d5ba37b-kube-api-access-4xc62\") pod \"multus-additional-cni-plugins-qclgb\" (UID: \"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\") " pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.911452 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ghr7\" (UniqueName: \"kubernetes.io/projected/7677a2dc-6b77-4fbc-90e3-9548a2298016-kube-api-access-5ghr7\") pod \"node-resolver-dqg6t\" (UID: \"7677a2dc-6b77-4fbc-90e3-9548a2298016\") " pod="openshift-dns/node-resolver-dqg6t" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.912029 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md8dh\" (UniqueName: \"kubernetes.io/projected/97e822d6-58fe-41f4-b08b-3c9b42273307-kube-api-access-md8dh\") pod \"multus-hmbvn\" (UID: \"97e822d6-58fe-41f4-b08b-3c9b42273307\") " pod="openshift-multus/multus-hmbvn" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.921849 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.938459 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.956416 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.970902 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:18 crc kubenswrapper[4802]: I1125 16:47:18.985214 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:18Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.001987 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-dqg6t" Nov 25 16:47:19 crc kubenswrapper[4802]: W1125 16:47:19.014552 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7677a2dc_6b77_4fbc_90e3_9548a2298016.slice/crio-08f5972c42d8e7024a597c50447a3ca9f849263a3606eeade1bee04c37d40542 WatchSource:0}: Error finding container 08f5972c42d8e7024a597c50447a3ca9f849263a3606eeade1bee04c37d40542: Status 404 returned error can't find the container with id 08f5972c42d8e7024a597c50447a3ca9f849263a3606eeade1bee04c37d40542 Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.019539 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-hmbvn" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.029842 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 16:47:19 crc kubenswrapper[4802]: W1125 16:47:19.031260 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97e822d6_58fe_41f4_b08b_3c9b42273307.slice/crio-d8e7c0b9a05033c2908787edf790792259ca0121c00eb485f4e2c3e4c82689e2 WatchSource:0}: Error finding container d8e7c0b9a05033c2908787edf790792259ca0121c00eb485f4e2c3e4c82689e2: Status 404 returned error can't find the container with id d8e7c0b9a05033c2908787edf790792259ca0121c00eb485f4e2c3e4c82689e2 Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.038228 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-qclgb" Nov 25 16:47:19 crc kubenswrapper[4802]: W1125 16:47:19.042353 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c3e23e7_2aae_4a81_b64f_4ed182a43d5e.slice/crio-035c73a7844348c1cf909b598622a8ef360576ee046629274ddb86970d589e6f WatchSource:0}: Error finding container 035c73a7844348c1cf909b598622a8ef360576ee046629274ddb86970d589e6f: Status 404 returned error can't find the container with id 035c73a7844348c1cf909b598622a8ef360576ee046629274ddb86970d589e6f Nov 25 16:47:19 crc kubenswrapper[4802]: W1125 16:47:19.056672 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32628afd_a82e_48a8_a3d2_cfb23d5ba37b.slice/crio-d3fb2b34d16ed3bb04165c435c9875d7e32db39bc40f8580ca19bf4c61c08f6f WatchSource:0}: Error finding container d3fb2b34d16ed3bb04165c435c9875d7e32db39bc40f8580ca19bf4c61c08f6f: Status 404 returned error can't find the container with id d3fb2b34d16ed3bb04165c435c9875d7e32db39bc40f8580ca19bf4c61c08f6f Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.078710 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dkxhj"] Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.079687 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.081929 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.082015 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.082182 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.082304 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.082465 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.082593 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.082713 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.110343 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.157060 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.173839 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.186286 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.192566 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-log-socket\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.192622 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-systemd-units\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.192645 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-slash\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.192666 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-run-netns\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.192693 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bad5c073-f5d9-4410-9350-bb2a51a764a2-ovnkube-script-lib\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.192716 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bad5c073-f5d9-4410-9350-bb2a51a764a2-ovnkube-config\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.192738 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-var-lib-openvswitch\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.192770 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-kubelet\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.192789 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-run-systemd\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.192809 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-etc-openvswitch\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.192829 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-node-log\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.192850 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-cni-netd\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.192872 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-run-openvswitch\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.192917 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.192994 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bad5c073-f5d9-4410-9350-bb2a51a764a2-env-overrides\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.193120 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-cni-bin\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.193216 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgb5h\" (UniqueName: \"kubernetes.io/projected/bad5c073-f5d9-4410-9350-bb2a51a764a2-kube-api-access-bgb5h\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.193248 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-run-ovn\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.193267 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bad5c073-f5d9-4410-9350-bb2a51a764a2-ovn-node-metrics-cert\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.193286 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-run-ovn-kubernetes\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.200180 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.220333 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.238727 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.251019 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.262636 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.275990 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294432 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-var-lib-openvswitch\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294490 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-kubelet\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294516 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-run-systemd\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294535 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-etc-openvswitch\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294555 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-node-log\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294576 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-cni-netd\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294614 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-run-openvswitch\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294636 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294656 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-run-systemd\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294700 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bad5c073-f5d9-4410-9350-bb2a51a764a2-env-overrides\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294722 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-cni-bin\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294733 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-var-lib-openvswitch\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294742 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-run-ovn\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294762 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgb5h\" (UniqueName: \"kubernetes.io/projected/bad5c073-f5d9-4410-9350-bb2a51a764a2-kube-api-access-bgb5h\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294769 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-kubelet\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294782 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-run-ovn-kubernetes\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294804 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bad5c073-f5d9-4410-9350-bb2a51a764a2-ovn-node-metrics-cert\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294809 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294825 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-log-socket\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294857 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-systemd-units\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294879 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-slash\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294886 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-etc-openvswitch\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294903 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-run-netns\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294928 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-node-log\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294934 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bad5c073-f5d9-4410-9350-bb2a51a764a2-ovnkube-script-lib\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.294990 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bad5c073-f5d9-4410-9350-bb2a51a764a2-ovnkube-config\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.295775 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bad5c073-f5d9-4410-9350-bb2a51a764a2-ovnkube-script-lib\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.295945 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bad5c073-f5d9-4410-9350-bb2a51a764a2-ovnkube-config\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.295993 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-cni-netd\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.296025 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-run-openvswitch\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.296201 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bad5c073-f5d9-4410-9350-bb2a51a764a2-env-overrides\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.296252 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-slash\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.296282 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-run-netns\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.296308 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-log-socket\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.296300 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-systemd-units\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.296337 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-run-ovn\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.296372 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-cni-bin\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.296402 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-run-ovn-kubernetes\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.301631 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.303998 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bad5c073-f5d9-4410-9350-bb2a51a764a2-ovn-node-metrics-cert\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.314707 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgb5h\" (UniqueName: \"kubernetes.io/projected/bad5c073-f5d9-4410-9350-bb2a51a764a2-kube-api-access-bgb5h\") pod \"ovnkube-node-dkxhj\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.334195 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.416600 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:19 crc kubenswrapper[4802]: W1125 16:47:19.427627 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbad5c073_f5d9_4410_9350_bb2a51a764a2.slice/crio-eaf11c74cfd2e012d33899947df8e98203e35baa2ed68906bf688745fbeaed6c WatchSource:0}: Error finding container eaf11c74cfd2e012d33899947df8e98203e35baa2ed68906bf688745fbeaed6c: Status 404 returned error can't find the container with id eaf11c74cfd2e012d33899947df8e98203e35baa2ed68906bf688745fbeaed6c Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.504142 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:19 crc kubenswrapper[4802]: E1125 16:47:19.504259 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.598724 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-dqg6t" event={"ID":"7677a2dc-6b77-4fbc-90e3-9548a2298016","Type":"ContainerStarted","Data":"f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa"} Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.598990 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-dqg6t" event={"ID":"7677a2dc-6b77-4fbc-90e3-9548a2298016","Type":"ContainerStarted","Data":"08f5972c42d8e7024a597c50447a3ca9f849263a3606eeade1bee04c37d40542"} Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.600624 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerStarted","Data":"f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23"} Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.600665 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerStarted","Data":"5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef"} Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.600675 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerStarted","Data":"035c73a7844348c1cf909b598622a8ef360576ee046629274ddb86970d589e6f"} Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.602264 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hmbvn" event={"ID":"97e822d6-58fe-41f4-b08b-3c9b42273307","Type":"ContainerStarted","Data":"ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b"} Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.602391 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hmbvn" event={"ID":"97e822d6-58fe-41f4-b08b-3c9b42273307","Type":"ContainerStarted","Data":"d8e7c0b9a05033c2908787edf790792259ca0121c00eb485f4e2c3e4c82689e2"} Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.603878 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097"} Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.608928 4802 generic.go:334] "Generic (PLEG): container finished" podID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerID="c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99" exitCode=0 Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.609031 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerDied","Data":"c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99"} Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.609062 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerStarted","Data":"eaf11c74cfd2e012d33899947df8e98203e35baa2ed68906bf688745fbeaed6c"} Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.612171 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" event={"ID":"32628afd-a82e-48a8-a3d2-cfb23d5ba37b","Type":"ContainerStarted","Data":"9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f"} Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.612205 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" event={"ID":"32628afd-a82e-48a8-a3d2-cfb23d5ba37b","Type":"ContainerStarted","Data":"d3fb2b34d16ed3bb04165c435c9875d7e32db39bc40f8580ca19bf4c61c08f6f"} Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.623782 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.635918 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.647616 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.660088 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.672006 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.686738 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.698997 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.710666 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.726718 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.740150 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.753188 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.766057 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.784264 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.797690 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.810675 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.824054 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.837088 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.848936 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.864607 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.892405 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.925853 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.940956 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.952243 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:19 crc kubenswrapper[4802]: I1125 16:47:19.966928 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:19Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.304762 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.304864 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.304901 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.304924 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:20 crc kubenswrapper[4802]: E1125 16:47:20.304974 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:47:24.304944522 +0000 UTC m=+27.449291718 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:47:20 crc kubenswrapper[4802]: E1125 16:47:20.305006 4802 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 16:47:20 crc kubenswrapper[4802]: E1125 16:47:20.305079 4802 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 16:47:20 crc kubenswrapper[4802]: E1125 16:47:20.305090 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:24.305071785 +0000 UTC m=+27.449418981 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 16:47:20 crc kubenswrapper[4802]: E1125 16:47:20.305089 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 16:47:20 crc kubenswrapper[4802]: E1125 16:47:20.305179 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 16:47:20 crc kubenswrapper[4802]: E1125 16:47:20.305190 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:24.305162577 +0000 UTC m=+27.449509823 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 16:47:20 crc kubenswrapper[4802]: E1125 16:47:20.305212 4802 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:20 crc kubenswrapper[4802]: E1125 16:47:20.305259 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:24.305247909 +0000 UTC m=+27.449595165 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.406374 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:20 crc kubenswrapper[4802]: E1125 16:47:20.406547 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 16:47:20 crc kubenswrapper[4802]: E1125 16:47:20.406567 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 16:47:20 crc kubenswrapper[4802]: E1125 16:47:20.406579 4802 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:20 crc kubenswrapper[4802]: E1125 16:47:20.406636 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:24.406615551 +0000 UTC m=+27.550962737 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.503531 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:20 crc kubenswrapper[4802]: E1125 16:47:20.503913 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.504247 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:20 crc kubenswrapper[4802]: E1125 16:47:20.504301 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.601763 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-2pbmn"] Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.602196 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2pbmn" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.603994 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.604197 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.604197 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.604358 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.618428 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerStarted","Data":"1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4"} Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.618462 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerStarted","Data":"857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542"} Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.618474 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerStarted","Data":"e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816"} Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.618485 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerStarted","Data":"a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3"} Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.618495 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerStarted","Data":"ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8"} Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.618504 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerStarted","Data":"90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8"} Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.619907 4802 generic.go:334] "Generic (PLEG): container finished" podID="32628afd-a82e-48a8-a3d2-cfb23d5ba37b" containerID="9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f" exitCode=0 Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.619926 4802 generic.go:334] "Generic (PLEG): container finished" podID="32628afd-a82e-48a8-a3d2-cfb23d5ba37b" containerID="a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a" exitCode=0 Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.620022 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" event={"ID":"32628afd-a82e-48a8-a3d2-cfb23d5ba37b","Type":"ContainerDied","Data":"9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f"} Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.620065 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" event={"ID":"32628afd-a82e-48a8-a3d2-cfb23d5ba37b","Type":"ContainerDied","Data":"a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a"} Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.621266 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:20Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.816214 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:20Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.843307 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:20Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.856296 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:20Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.867955 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:20Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.877603 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:20Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.893344 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:20Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.907110 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:20Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.910562 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9f6394dc-d7bd-4205-9ff4-aca84d1306b0-serviceca\") pod \"node-ca-2pbmn\" (UID: \"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\") " pod="openshift-image-registry/node-ca-2pbmn" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.910659 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn92s\" (UniqueName: \"kubernetes.io/projected/9f6394dc-d7bd-4205-9ff4-aca84d1306b0-kube-api-access-kn92s\") pod \"node-ca-2pbmn\" (UID: \"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\") " pod="openshift-image-registry/node-ca-2pbmn" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.910712 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f6394dc-d7bd-4205-9ff4-aca84d1306b0-host\") pod \"node-ca-2pbmn\" (UID: \"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\") " pod="openshift-image-registry/node-ca-2pbmn" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.919650 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:20Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.930474 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:20Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.942098 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:20Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.952750 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:20Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.970263 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:20Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.985332 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:20Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:20 crc kubenswrapper[4802]: I1125 16:47:20.996367 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:20Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.007331 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.011479 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9f6394dc-d7bd-4205-9ff4-aca84d1306b0-serviceca\") pod \"node-ca-2pbmn\" (UID: \"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\") " pod="openshift-image-registry/node-ca-2pbmn" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.011530 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn92s\" (UniqueName: \"kubernetes.io/projected/9f6394dc-d7bd-4205-9ff4-aca84d1306b0-kube-api-access-kn92s\") pod \"node-ca-2pbmn\" (UID: \"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\") " pod="openshift-image-registry/node-ca-2pbmn" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.011553 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f6394dc-d7bd-4205-9ff4-aca84d1306b0-host\") pod \"node-ca-2pbmn\" (UID: \"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\") " pod="openshift-image-registry/node-ca-2pbmn" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.011601 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f6394dc-d7bd-4205-9ff4-aca84d1306b0-host\") pod \"node-ca-2pbmn\" (UID: \"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\") " pod="openshift-image-registry/node-ca-2pbmn" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.012681 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9f6394dc-d7bd-4205-9ff4-aca84d1306b0-serviceca\") pod \"node-ca-2pbmn\" (UID: \"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\") " pod="openshift-image-registry/node-ca-2pbmn" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.019397 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.029662 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.031899 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn92s\" (UniqueName: \"kubernetes.io/projected/9f6394dc-d7bd-4205-9ff4-aca84d1306b0-kube-api-access-kn92s\") pod \"node-ca-2pbmn\" (UID: \"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\") " pod="openshift-image-registry/node-ca-2pbmn" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.046210 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.060777 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.080196 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.088817 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.102679 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.113384 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.114471 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2pbmn" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.125579 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: W1125 16:47:21.126198 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f6394dc_d7bd_4205_9ff4_aca84d1306b0.slice/crio-1a6a0fe1264da5292c45f47f7b263bde4badfd21cadb85a1d492e4e8d7c5d3fd WatchSource:0}: Error finding container 1a6a0fe1264da5292c45f47f7b263bde4badfd21cadb85a1d492e4e8d7c5d3fd: Status 404 returned error can't find the container with id 1a6a0fe1264da5292c45f47f7b263bde4badfd21cadb85a1d492e4e8d7c5d3fd Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.136306 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.503894 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:21 crc kubenswrapper[4802]: E1125 16:47:21.504377 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.623958 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2pbmn" event={"ID":"9f6394dc-d7bd-4205-9ff4-aca84d1306b0","Type":"ContainerStarted","Data":"b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde"} Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.624017 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2pbmn" event={"ID":"9f6394dc-d7bd-4205-9ff4-aca84d1306b0","Type":"ContainerStarted","Data":"1a6a0fe1264da5292c45f47f7b263bde4badfd21cadb85a1d492e4e8d7c5d3fd"} Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.627541 4802 generic.go:334] "Generic (PLEG): container finished" podID="32628afd-a82e-48a8-a3d2-cfb23d5ba37b" containerID="f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb" exitCode=0 Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.627616 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" event={"ID":"32628afd-a82e-48a8-a3d2-cfb23d5ba37b","Type":"ContainerDied","Data":"f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb"} Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.646193 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.657961 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.669858 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.682056 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.696051 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.705420 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.721810 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.736132 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.749949 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.761993 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.784226 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.796143 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.809878 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.822394 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.840160 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.851455 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.864313 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.877497 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.889325 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.901689 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.912410 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.925579 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.938490 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.951763 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.964447 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:21 crc kubenswrapper[4802]: I1125 16:47:21.978114 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:22 crc kubenswrapper[4802]: I1125 16:47:22.503824 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:22 crc kubenswrapper[4802]: I1125 16:47:22.503825 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:22 crc kubenswrapper[4802]: E1125 16:47:22.503950 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:22 crc kubenswrapper[4802]: E1125 16:47:22.504015 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:22 crc kubenswrapper[4802]: I1125 16:47:22.635025 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerStarted","Data":"8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb"} Nov 25 16:47:22 crc kubenswrapper[4802]: I1125 16:47:22.637689 4802 generic.go:334] "Generic (PLEG): container finished" podID="32628afd-a82e-48a8-a3d2-cfb23d5ba37b" containerID="c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94" exitCode=0 Nov 25 16:47:22 crc kubenswrapper[4802]: I1125 16:47:22.637740 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" event={"ID":"32628afd-a82e-48a8-a3d2-cfb23d5ba37b","Type":"ContainerDied","Data":"c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94"} Nov 25 16:47:22 crc kubenswrapper[4802]: I1125 16:47:22.663963 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:22Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:22 crc kubenswrapper[4802]: I1125 16:47:22.678260 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:22Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:22 crc kubenswrapper[4802]: I1125 16:47:22.692061 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:22Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:22 crc kubenswrapper[4802]: I1125 16:47:22.707493 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:22Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:22 crc kubenswrapper[4802]: I1125 16:47:22.721816 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:22Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:22 crc kubenswrapper[4802]: I1125 16:47:22.733449 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:22Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:22 crc kubenswrapper[4802]: I1125 16:47:22.750825 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:22Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:22 crc kubenswrapper[4802]: I1125 16:47:22.766274 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:22Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:22 crc kubenswrapper[4802]: I1125 16:47:22.785521 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:22Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:22 crc kubenswrapper[4802]: I1125 16:47:22.797517 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:22Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:22 crc kubenswrapper[4802]: I1125 16:47:22.810015 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:22Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:22 crc kubenswrapper[4802]: I1125 16:47:22.825095 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:22Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:22 crc kubenswrapper[4802]: I1125 16:47:22.837602 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:22Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.023188 4802 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.024783 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.024815 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.024823 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.024909 4802 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.030616 4802 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.030791 4802 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.031792 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.031829 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.031838 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.031853 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.031865 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:23Z","lastTransitionTime":"2025-11-25T16:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:23 crc kubenswrapper[4802]: E1125 16:47:23.043356 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.046920 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.046967 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.046979 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.046996 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.047008 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:23Z","lastTransitionTime":"2025-11-25T16:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:23 crc kubenswrapper[4802]: E1125 16:47:23.058293 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.062086 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.062138 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.062152 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.062173 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.062188 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:23Z","lastTransitionTime":"2025-11-25T16:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:23 crc kubenswrapper[4802]: E1125 16:47:23.074304 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.077348 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.077412 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.077425 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.077442 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.077453 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:23Z","lastTransitionTime":"2025-11-25T16:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:23 crc kubenswrapper[4802]: E1125 16:47:23.087168 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.090369 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.090407 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.090419 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.090435 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.090448 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:23Z","lastTransitionTime":"2025-11-25T16:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:23 crc kubenswrapper[4802]: E1125 16:47:23.102945 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: E1125 16:47:23.103148 4802 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.105264 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.105303 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.105327 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.105346 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.105357 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:23Z","lastTransitionTime":"2025-11-25T16:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.207864 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.207900 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.207909 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.207923 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.207934 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:23Z","lastTransitionTime":"2025-11-25T16:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.275118 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.284902 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.286706 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.288053 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.300583 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.309774 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.309818 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.309826 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.309841 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.309851 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:23Z","lastTransitionTime":"2025-11-25T16:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.312840 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.324397 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.337317 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.349657 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.365135 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.382390 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.393466 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.410316 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.412665 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.412881 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.413239 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.413364 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.413465 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:23Z","lastTransitionTime":"2025-11-25T16:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.422035 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.433694 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.444541 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.457259 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.472068 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.485239 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.495352 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.504539 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:23 crc kubenswrapper[4802]: E1125 16:47:23.504711 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.510720 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.516255 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.516380 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.516438 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.516513 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.516604 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:23Z","lastTransitionTime":"2025-11-25T16:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.522963 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.533285 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.548001 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.559021 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.574091 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.586965 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.597410 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.615405 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.618599 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.618637 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.618647 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.618661 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.618671 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:23Z","lastTransitionTime":"2025-11-25T16:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.625085 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.643134 4802 generic.go:334] "Generic (PLEG): container finished" podID="32628afd-a82e-48a8-a3d2-cfb23d5ba37b" containerID="f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d" exitCode=0 Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.643659 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" event={"ID":"32628afd-a82e-48a8-a3d2-cfb23d5ba37b","Type":"ContainerDied","Data":"f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d"} Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.663103 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.675515 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.687333 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.699748 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.714688 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.721666 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.721704 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.721719 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.721739 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.721752 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:23Z","lastTransitionTime":"2025-11-25T16:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.724570 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.735175 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.745834 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.756833 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.770309 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.782715 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.794986 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.807334 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.823810 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.823860 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.823870 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.823891 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.824050 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:23Z","lastTransitionTime":"2025-11-25T16:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.836201 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:23Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.926647 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.926713 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.926722 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.926736 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:23 crc kubenswrapper[4802]: I1125 16:47:23.926747 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:23Z","lastTransitionTime":"2025-11-25T16:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.028566 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.028624 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.028642 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.028662 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.028674 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:24Z","lastTransitionTime":"2025-11-25T16:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.130931 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.131167 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.131266 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.131335 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.131391 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:24Z","lastTransitionTime":"2025-11-25T16:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.233747 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.233785 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.233796 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.233815 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.233826 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:24Z","lastTransitionTime":"2025-11-25T16:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.337260 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.337308 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.337320 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.337338 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.337353 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:24Z","lastTransitionTime":"2025-11-25T16:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.345212 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.345401 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:24 crc kubenswrapper[4802]: E1125 16:47:24.345568 4802 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 16:47:24 crc kubenswrapper[4802]: E1125 16:47:24.345660 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:32.34563456 +0000 UTC m=+35.489981766 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.346088 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:24 crc kubenswrapper[4802]: E1125 16:47:24.346174 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:47:32.346155683 +0000 UTC m=+35.490502949 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.346198 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:24 crc kubenswrapper[4802]: E1125 16:47:24.346379 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 16:47:24 crc kubenswrapper[4802]: E1125 16:47:24.346394 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 16:47:24 crc kubenswrapper[4802]: E1125 16:47:24.346410 4802 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:24 crc kubenswrapper[4802]: E1125 16:47:24.346398 4802 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 16:47:24 crc kubenswrapper[4802]: E1125 16:47:24.346489 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:32.346475302 +0000 UTC m=+35.490822488 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:24 crc kubenswrapper[4802]: E1125 16:47:24.346603 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:32.346591894 +0000 UTC m=+35.490939080 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.440201 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.440241 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.440253 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.440269 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.440279 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:24Z","lastTransitionTime":"2025-11-25T16:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.447376 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:24 crc kubenswrapper[4802]: E1125 16:47:24.447623 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 16:47:24 crc kubenswrapper[4802]: E1125 16:47:24.447664 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 16:47:24 crc kubenswrapper[4802]: E1125 16:47:24.447683 4802 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:24 crc kubenswrapper[4802]: E1125 16:47:24.447761 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:32.44774074 +0000 UTC m=+35.592087926 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.503595 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.503603 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:24 crc kubenswrapper[4802]: E1125 16:47:24.503779 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:24 crc kubenswrapper[4802]: E1125 16:47:24.503847 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.542031 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.542063 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.542073 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.542088 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.542099 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:24Z","lastTransitionTime":"2025-11-25T16:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.644360 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.645264 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.645291 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.645313 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.645334 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:24Z","lastTransitionTime":"2025-11-25T16:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.649500 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerStarted","Data":"001fe8eaa8178997a0bc80ce5bdc6eb4ec52b5fbc1b4d46c3c0140e48b110a55"} Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.649695 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.652869 4802 generic.go:334] "Generic (PLEG): container finished" podID="32628afd-a82e-48a8-a3d2-cfb23d5ba37b" containerID="9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8" exitCode=0 Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.652899 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" event={"ID":"32628afd-a82e-48a8-a3d2-cfb23d5ba37b","Type":"ContainerDied","Data":"9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8"} Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.671106 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001fe8eaa8178997a0bc80ce5bdc6eb4ec52b5fbc1b4d46c3c0140e48b110a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.675371 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.681877 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.697043 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.710852 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.724182 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.734420 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.748157 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.748194 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.748206 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.748220 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.748232 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:24Z","lastTransitionTime":"2025-11-25T16:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.749582 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.760725 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.776328 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.791877 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.804665 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.819662 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.831861 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.847348 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.853350 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.853387 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.853397 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.853411 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.853422 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:24Z","lastTransitionTime":"2025-11-25T16:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.857766 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.877556 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001fe8eaa8178997a0bc80ce5bdc6eb4ec52b5fbc1b4d46c3c0140e48b110a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.895243 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.911828 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.925316 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.944387 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.956806 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.956857 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.956871 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.956889 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.956900 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:24Z","lastTransitionTime":"2025-11-25T16:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.957887 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.970086 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.982186 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:24 crc kubenswrapper[4802]: I1125 16:47:24.993936 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.008937 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.027289 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.044025 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.057740 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.059306 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.059344 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.059358 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.059377 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.059391 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:25Z","lastTransitionTime":"2025-11-25T16:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.162476 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.162513 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.162524 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.162541 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.162552 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:25Z","lastTransitionTime":"2025-11-25T16:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.264793 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.264824 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.264833 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.264846 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.264855 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:25Z","lastTransitionTime":"2025-11-25T16:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.368819 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.368881 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.368898 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.368925 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.368941 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:25Z","lastTransitionTime":"2025-11-25T16:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.472227 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.472311 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.472333 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.472369 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.472390 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:25Z","lastTransitionTime":"2025-11-25T16:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.503644 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:25 crc kubenswrapper[4802]: E1125 16:47:25.503830 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.575638 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.575675 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.575684 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.575702 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.575712 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:25Z","lastTransitionTime":"2025-11-25T16:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.661765 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" event={"ID":"32628afd-a82e-48a8-a3d2-cfb23d5ba37b","Type":"ContainerStarted","Data":"3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862"} Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.661911 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.662747 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.679382 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.679844 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.679859 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.679884 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.679899 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:25Z","lastTransitionTime":"2025-11-25T16:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.681280 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.695259 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.695287 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.710791 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.722485 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.734300 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.748140 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.760206 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.772474 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.782652 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.782686 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.782694 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.782709 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.782721 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:25Z","lastTransitionTime":"2025-11-25T16:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.785438 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.800957 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.819142 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.833265 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.852386 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001fe8eaa8178997a0bc80ce5bdc6eb4ec52b5fbc1b4d46c3c0140e48b110a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.861647 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.873207 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.883094 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.884758 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.884792 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.884801 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.884817 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.884829 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:25Z","lastTransitionTime":"2025-11-25T16:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.896330 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.909490 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.924022 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.939864 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.952172 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.967736 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.979624 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.987807 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.987852 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.987864 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.987885 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.987899 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:25Z","lastTransitionTime":"2025-11-25T16:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:25 crc kubenswrapper[4802]: I1125 16:47:25.998295 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001fe8eaa8178997a0bc80ce5bdc6eb4ec52b5fbc1b4d46c3c0140e48b110a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:25Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.008470 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:26Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.022293 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:26Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.037254 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:26Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.077020 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:26Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.091405 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.091461 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.091471 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.091493 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.091507 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:26Z","lastTransitionTime":"2025-11-25T16:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.194617 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.194666 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.194678 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.194697 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.194710 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:26Z","lastTransitionTime":"2025-11-25T16:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.298418 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.298470 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.298483 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.298506 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.298575 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:26Z","lastTransitionTime":"2025-11-25T16:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.401825 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.401864 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.401873 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.401889 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.401902 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:26Z","lastTransitionTime":"2025-11-25T16:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.503954 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.504017 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:26 crc kubenswrapper[4802]: E1125 16:47:26.504083 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:26 crc kubenswrapper[4802]: E1125 16:47:26.504277 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.504833 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.504865 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.504878 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.504893 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.504905 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:26Z","lastTransitionTime":"2025-11-25T16:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.607732 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.607795 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.607804 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.607818 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.607829 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:26Z","lastTransitionTime":"2025-11-25T16:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.664846 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.710435 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.710464 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.710472 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.710484 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.710493 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:26Z","lastTransitionTime":"2025-11-25T16:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.813266 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.813310 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.813320 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.813339 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.813351 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:26Z","lastTransitionTime":"2025-11-25T16:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.915419 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.915458 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.915470 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.915486 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:26 crc kubenswrapper[4802]: I1125 16:47:26.915499 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:26Z","lastTransitionTime":"2025-11-25T16:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.018062 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.018099 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.018110 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.018144 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.018165 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:27Z","lastTransitionTime":"2025-11-25T16:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.120951 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.120996 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.121006 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.121021 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.121030 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:27Z","lastTransitionTime":"2025-11-25T16:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.223026 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.223073 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.223088 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.223104 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.223133 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:27Z","lastTransitionTime":"2025-11-25T16:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.325554 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.325592 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.325610 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.325628 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.325641 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:27Z","lastTransitionTime":"2025-11-25T16:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.428183 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.428220 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.428231 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.428245 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.428255 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:27Z","lastTransitionTime":"2025-11-25T16:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.504407 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:27 crc kubenswrapper[4802]: E1125 16:47:27.504601 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.522851 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001fe8eaa8178997a0bc80ce5bdc6eb4ec52b5fbc1b4d46c3c0140e48b110a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.530473 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.530508 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.530519 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.530536 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.530547 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:27Z","lastTransitionTime":"2025-11-25T16:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.531464 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.542315 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.553510 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.564879 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.574804 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.590573 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.603295 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.615851 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.628960 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.632289 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.632326 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.632335 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.632349 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.632359 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:27Z","lastTransitionTime":"2025-11-25T16:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.641726 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.651889 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.664591 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.669272 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovnkube-controller/0.log" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.671529 4802 generic.go:334] "Generic (PLEG): container finished" podID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerID="001fe8eaa8178997a0bc80ce5bdc6eb4ec52b5fbc1b4d46c3c0140e48b110a55" exitCode=1 Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.671568 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerDied","Data":"001fe8eaa8178997a0bc80ce5bdc6eb4ec52b5fbc1b4d46c3c0140e48b110a55"} Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.672142 4802 scope.go:117] "RemoveContainer" containerID="001fe8eaa8178997a0bc80ce5bdc6eb4ec52b5fbc1b4d46c3c0140e48b110a55" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.677703 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.696551 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001fe8eaa8178997a0bc80ce5bdc6eb4ec52b5fbc1b4d46c3c0140e48b110a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001fe8eaa8178997a0bc80ce5bdc6eb4ec52b5fbc1b4d46c3c0140e48b110a55\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:27Z\\\",\\\"message\\\":\\\"or *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 16:47:27.398540 6043 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 16:47:27.398104 6043 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 16:47:27.398603 6043 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 16:47:27.398612 6043 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 16:47:27.398631 6043 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 16:47:27.398644 6043 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 16:47:27.398656 6043 factory.go:656] Stopping watch factory\\\\nI1125 16:47:27.398668 6043 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 16:47:27.398675 6043 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 16:47:27.398681 6043 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 16:47:27.398686 6043 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 16:47:27.398692 6043 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 16:47:27.398704 6043 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 16:47:27.398145 6043 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.705920 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.717453 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.728904 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.734372 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.734396 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.734405 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.734421 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.734614 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:27Z","lastTransitionTime":"2025-11-25T16:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.741816 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.755320 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.769878 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.783444 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.796595 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.811443 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.827086 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.836827 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.836863 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.836871 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.836886 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.836899 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:27Z","lastTransitionTime":"2025-11-25T16:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.839506 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.858618 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.875316 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.939401 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.939437 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.939454 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.939471 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:27 crc kubenswrapper[4802]: I1125 16:47:27.939481 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:27Z","lastTransitionTime":"2025-11-25T16:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.041642 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.041680 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.041691 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.041707 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.041718 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:28Z","lastTransitionTime":"2025-11-25T16:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.144789 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.144844 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.144856 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.144876 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.144890 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:28Z","lastTransitionTime":"2025-11-25T16:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.248010 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.248061 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.248075 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.248097 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.248107 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:28Z","lastTransitionTime":"2025-11-25T16:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.351478 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.351523 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.351536 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.351556 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.351567 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:28Z","lastTransitionTime":"2025-11-25T16:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.454331 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.454386 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.454396 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.454413 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.454422 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:28Z","lastTransitionTime":"2025-11-25T16:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.504654 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.504669 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:28 crc kubenswrapper[4802]: E1125 16:47:28.504887 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:28 crc kubenswrapper[4802]: E1125 16:47:28.505256 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.557326 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.557386 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.557398 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.557422 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.557436 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:28Z","lastTransitionTime":"2025-11-25T16:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.659944 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.659978 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.659986 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.660000 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.660009 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:28Z","lastTransitionTime":"2025-11-25T16:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.676208 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovnkube-controller/1.log" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.677290 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovnkube-controller/0.log" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.680363 4802 generic.go:334] "Generic (PLEG): container finished" podID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerID="eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa" exitCode=1 Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.680425 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerDied","Data":"eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa"} Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.680468 4802 scope.go:117] "RemoveContainer" containerID="001fe8eaa8178997a0bc80ce5bdc6eb4ec52b5fbc1b4d46c3c0140e48b110a55" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.681093 4802 scope.go:117] "RemoveContainer" containerID="eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa" Nov 25 16:47:28 crc kubenswrapper[4802]: E1125 16:47:28.681245 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.695015 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.713162 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001fe8eaa8178997a0bc80ce5bdc6eb4ec52b5fbc1b4d46c3c0140e48b110a55\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:27Z\\\",\\\"message\\\":\\\"or *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 16:47:27.398540 6043 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 16:47:27.398104 6043 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 16:47:27.398603 6043 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 16:47:27.398612 6043 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 16:47:27.398631 6043 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 16:47:27.398644 6043 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 16:47:27.398656 6043 factory.go:656] Stopping watch factory\\\\nI1125 16:47:27.398668 6043 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 16:47:27.398675 6043 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 16:47:27.398681 6043 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 16:47:27.398686 6043 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 16:47:27.398692 6043 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 16:47:27.398704 6043 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 16:47:27.398145 6043 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:28Z\\\",\\\"message\\\":\\\"ce\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}\\\\nF1125 16:47:28.617785 6189 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z]\\\\nI1125 16:47:28.617784 6189 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-service-ca-operator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid ==\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.725856 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.739228 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.752639 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.762653 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.763221 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.763321 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.763423 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.763503 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:28Z","lastTransitionTime":"2025-11-25T16:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.765965 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.777882 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.790507 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.805990 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.820779 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.834796 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.846662 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.861237 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.865536 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.865572 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.865580 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.865594 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.865606 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:28Z","lastTransitionTime":"2025-11-25T16:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.874360 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.967992 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.968038 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.968051 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.968071 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:28 crc kubenswrapper[4802]: I1125 16:47:28.968083 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:28Z","lastTransitionTime":"2025-11-25T16:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.071305 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.071354 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.071368 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.071387 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.071398 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:29Z","lastTransitionTime":"2025-11-25T16:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.174908 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.174974 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.174988 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.175009 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.175022 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:29Z","lastTransitionTime":"2025-11-25T16:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.278694 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.278750 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.278763 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.278781 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.278795 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:29Z","lastTransitionTime":"2025-11-25T16:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.381855 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.381917 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.381929 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.381952 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.381967 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:29Z","lastTransitionTime":"2025-11-25T16:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.485098 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.485176 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.485192 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.485211 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.485223 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:29Z","lastTransitionTime":"2025-11-25T16:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.504319 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:29 crc kubenswrapper[4802]: E1125 16:47:29.504430 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.587113 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.587159 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.587166 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.587179 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.587188 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:29Z","lastTransitionTime":"2025-11-25T16:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.686285 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovnkube-controller/1.log" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.689231 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.689266 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.689276 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.689297 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.689308 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:29Z","lastTransitionTime":"2025-11-25T16:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.691619 4802 scope.go:117] "RemoveContainer" containerID="eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa" Nov 25 16:47:29 crc kubenswrapper[4802]: E1125 16:47:29.691828 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.706055 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.715380 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:28Z\\\",\\\"message\\\":\\\"ce\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}\\\\nF1125 16:47:28.617785 6189 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z]\\\\nI1125 16:47:28.617784 6189 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-service-ca-operator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid ==\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.729589 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.746649 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.759753 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.771691 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.785433 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.791607 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.791642 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.791651 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.791670 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.791688 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:29Z","lastTransitionTime":"2025-11-25T16:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.800742 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.815807 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.828267 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.840277 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.853680 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.869072 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.882277 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.894174 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.894217 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.894229 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.894252 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.894266 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:29Z","lastTransitionTime":"2025-11-25T16:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.898314 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.997910 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.997990 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.998009 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.998038 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:29 crc kubenswrapper[4802]: I1125 16:47:29.998055 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:29Z","lastTransitionTime":"2025-11-25T16:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.100574 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.100633 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.100641 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.100655 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.100664 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:30Z","lastTransitionTime":"2025-11-25T16:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.203154 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.203218 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.203232 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.203259 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.203272 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:30Z","lastTransitionTime":"2025-11-25T16:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.305743 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.305791 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.305803 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.305822 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.305835 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:30Z","lastTransitionTime":"2025-11-25T16:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.408975 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.409019 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.409028 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.409045 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.409055 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:30Z","lastTransitionTime":"2025-11-25T16:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.504018 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.504033 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:30 crc kubenswrapper[4802]: E1125 16:47:30.504179 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:30 crc kubenswrapper[4802]: E1125 16:47:30.504246 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.510918 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.510955 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.510968 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.510982 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.510994 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:30Z","lastTransitionTime":"2025-11-25T16:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.613672 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.613703 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.613713 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.613725 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.613735 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:30Z","lastTransitionTime":"2025-11-25T16:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.693954 4802 scope.go:117] "RemoveContainer" containerID="eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa" Nov 25 16:47:30 crc kubenswrapper[4802]: E1125 16:47:30.694136 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.716273 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.716337 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.716351 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.716379 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.716396 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:30Z","lastTransitionTime":"2025-11-25T16:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.818427 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.818472 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.818483 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.818500 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.818509 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:30Z","lastTransitionTime":"2025-11-25T16:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.920896 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.920937 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.920952 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.920968 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:30 crc kubenswrapper[4802]: I1125 16:47:30.920979 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:30Z","lastTransitionTime":"2025-11-25T16:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.023702 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.023739 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.023751 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.023768 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.023778 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:31Z","lastTransitionTime":"2025-11-25T16:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.126861 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.126942 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.126959 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.126986 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.127003 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:31Z","lastTransitionTime":"2025-11-25T16:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.153677 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh"] Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.154088 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.156096 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.156616 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.168459 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feaf869b-f88a-4874-8174-b689fcf6f40a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7p6zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.194299 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:28Z\\\",\\\"message\\\":\\\"ce\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}\\\\nF1125 16:47:28.617785 6189 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z]\\\\nI1125 16:47:28.617784 6189 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-service-ca-operator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid ==\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.206906 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.213196 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/feaf869b-f88a-4874-8174-b689fcf6f40a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7p6zh\" (UID: \"feaf869b-f88a-4874-8174-b689fcf6f40a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.213252 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/feaf869b-f88a-4874-8174-b689fcf6f40a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7p6zh\" (UID: \"feaf869b-f88a-4874-8174-b689fcf6f40a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.213314 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/feaf869b-f88a-4874-8174-b689fcf6f40a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7p6zh\" (UID: \"feaf869b-f88a-4874-8174-b689fcf6f40a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.213352 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqsfn\" (UniqueName: \"kubernetes.io/projected/feaf869b-f88a-4874-8174-b689fcf6f40a-kube-api-access-tqsfn\") pod \"ovnkube-control-plane-749d76644c-7p6zh\" (UID: \"feaf869b-f88a-4874-8174-b689fcf6f40a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.221799 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.229964 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.230014 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.230025 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.230049 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.230065 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:31Z","lastTransitionTime":"2025-11-25T16:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.239563 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.252667 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.263232 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.275706 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.288939 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.303744 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.314332 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/feaf869b-f88a-4874-8174-b689fcf6f40a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7p6zh\" (UID: \"feaf869b-f88a-4874-8174-b689fcf6f40a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.314375 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/feaf869b-f88a-4874-8174-b689fcf6f40a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7p6zh\" (UID: \"feaf869b-f88a-4874-8174-b689fcf6f40a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.314406 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/feaf869b-f88a-4874-8174-b689fcf6f40a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7p6zh\" (UID: \"feaf869b-f88a-4874-8174-b689fcf6f40a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.314449 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqsfn\" (UniqueName: \"kubernetes.io/projected/feaf869b-f88a-4874-8174-b689fcf6f40a-kube-api-access-tqsfn\") pod \"ovnkube-control-plane-749d76644c-7p6zh\" (UID: \"feaf869b-f88a-4874-8174-b689fcf6f40a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.315530 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/feaf869b-f88a-4874-8174-b689fcf6f40a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7p6zh\" (UID: \"feaf869b-f88a-4874-8174-b689fcf6f40a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.315602 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/feaf869b-f88a-4874-8174-b689fcf6f40a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7p6zh\" (UID: \"feaf869b-f88a-4874-8174-b689fcf6f40a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.318322 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.321922 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/feaf869b-f88a-4874-8174-b689fcf6f40a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7p6zh\" (UID: \"feaf869b-f88a-4874-8174-b689fcf6f40a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.330198 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqsfn\" (UniqueName: \"kubernetes.io/projected/feaf869b-f88a-4874-8174-b689fcf6f40a-kube-api-access-tqsfn\") pod \"ovnkube-control-plane-749d76644c-7p6zh\" (UID: \"feaf869b-f88a-4874-8174-b689fcf6f40a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.332728 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.332776 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.332789 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.332810 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.332822 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:31Z","lastTransitionTime":"2025-11-25T16:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.333507 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.347203 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.360033 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.374390 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.436238 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.436270 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.436278 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.436292 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.436301 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:31Z","lastTransitionTime":"2025-11-25T16:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.472334 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" Nov 25 16:47:31 crc kubenswrapper[4802]: W1125 16:47:31.488165 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeaf869b_f88a_4874_8174_b689fcf6f40a.slice/crio-f6f266a632a23218514171d9b0c237a4204eb8beb2787a8c7578fab032d6f3a8 WatchSource:0}: Error finding container f6f266a632a23218514171d9b0c237a4204eb8beb2787a8c7578fab032d6f3a8: Status 404 returned error can't find the container with id f6f266a632a23218514171d9b0c237a4204eb8beb2787a8c7578fab032d6f3a8 Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.504366 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:31 crc kubenswrapper[4802]: E1125 16:47:31.504539 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.539018 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.539554 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.539702 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.539763 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.539790 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:31Z","lastTransitionTime":"2025-11-25T16:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.643107 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.643164 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.643173 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.643190 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.643199 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:31Z","lastTransitionTime":"2025-11-25T16:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.697102 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" event={"ID":"feaf869b-f88a-4874-8174-b689fcf6f40a","Type":"ContainerStarted","Data":"d672b4cc7df9d1855c8e882c5f4c14dec0df01a8eadab587920753fee67abf33"} Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.697171 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" event={"ID":"feaf869b-f88a-4874-8174-b689fcf6f40a","Type":"ContainerStarted","Data":"f6f266a632a23218514171d9b0c237a4204eb8beb2787a8c7578fab032d6f3a8"} Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.746289 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.746743 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.746753 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.746773 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.746784 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:31Z","lastTransitionTime":"2025-11-25T16:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.849426 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.849474 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.849484 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.849503 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.849514 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:31Z","lastTransitionTime":"2025-11-25T16:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.953776 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.953814 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.953826 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.953841 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:31 crc kubenswrapper[4802]: I1125 16:47:31.953853 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:31Z","lastTransitionTime":"2025-11-25T16:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.055904 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.055946 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.055958 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.055973 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.055982 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:32Z","lastTransitionTime":"2025-11-25T16:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.158341 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.158382 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.158391 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.158404 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.158415 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:32Z","lastTransitionTime":"2025-11-25T16:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.261054 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.261091 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.261100 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.261116 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.261143 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:32Z","lastTransitionTime":"2025-11-25T16:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.362945 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.362985 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.363000 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.363017 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.363029 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:32Z","lastTransitionTime":"2025-11-25T16:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.425917 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.426014 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:32 crc kubenswrapper[4802]: E1125 16:47:32.426061 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:47:48.426034531 +0000 UTC m=+51.570381707 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:47:32 crc kubenswrapper[4802]: E1125 16:47:32.426157 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 16:47:32 crc kubenswrapper[4802]: E1125 16:47:32.426176 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.426174 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:32 crc kubenswrapper[4802]: E1125 16:47:32.426187 4802 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:32 crc kubenswrapper[4802]: E1125 16:47:32.426231 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:48.426218226 +0000 UTC m=+51.570565412 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.426212 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:32 crc kubenswrapper[4802]: E1125 16:47:32.426317 4802 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 16:47:32 crc kubenswrapper[4802]: E1125 16:47:32.426345 4802 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 16:47:32 crc kubenswrapper[4802]: E1125 16:47:32.426351 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:48.426344049 +0000 UTC m=+51.570691235 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 16:47:32 crc kubenswrapper[4802]: E1125 16:47:32.426396 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:48.42638381 +0000 UTC m=+51.570730986 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.465193 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.465233 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.465241 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.465255 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.465265 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:32Z","lastTransitionTime":"2025-11-25T16:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.503945 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.504034 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:32 crc kubenswrapper[4802]: E1125 16:47:32.504095 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:32 crc kubenswrapper[4802]: E1125 16:47:32.504188 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.527053 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:32 crc kubenswrapper[4802]: E1125 16:47:32.527233 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 16:47:32 crc kubenswrapper[4802]: E1125 16:47:32.527270 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 16:47:32 crc kubenswrapper[4802]: E1125 16:47:32.527282 4802 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:32 crc kubenswrapper[4802]: E1125 16:47:32.527343 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:48.527327671 +0000 UTC m=+51.671674857 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.567378 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.567417 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.567426 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.567440 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.567449 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:32Z","lastTransitionTime":"2025-11-25T16:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.610550 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-6ft2z"] Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.611037 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:32 crc kubenswrapper[4802]: E1125 16:47:32.611105 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.632032 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:28Z\\\",\\\"message\\\":\\\"ce\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}\\\\nF1125 16:47:28.617785 6189 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z]\\\\nI1125 16:47:28.617784 6189 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-service-ca-operator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid ==\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.640980 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.651890 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.663463 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.669999 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.670039 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.670051 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.670069 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.670080 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:32Z","lastTransitionTime":"2025-11-25T16:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.675637 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.685199 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.697395 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.701037 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" event={"ID":"feaf869b-f88a-4874-8174-b689fcf6f40a","Type":"ContainerStarted","Data":"92fbcf119fd704d01bca613b32439360422fe8ea79ba3331eace6d7f52cb5d5a"} Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.712841 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.728518 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.729193 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs\") pod \"network-metrics-daemon-6ft2z\" (UID: \"cd9aa596-76c2-468c-b732-d85bc1f16070\") " pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.729338 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q78q2\" (UniqueName: \"kubernetes.io/projected/cd9aa596-76c2-468c-b732-d85bc1f16070-kube-api-access-q78q2\") pod \"network-metrics-daemon-6ft2z\" (UID: \"cd9aa596-76c2-468c-b732-d85bc1f16070\") " pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.741180 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.753324 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.766408 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.772095 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.772146 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.772159 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.772183 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.772194 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:32Z","lastTransitionTime":"2025-11-25T16:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.778740 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.789515 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.798826 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd9aa596-76c2-468c-b732-d85bc1f16070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6ft2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.809262 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feaf869b-f88a-4874-8174-b689fcf6f40a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7p6zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.823673 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feaf869b-f88a-4874-8174-b689fcf6f40a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d672b4cc7df9d1855c8e882c5f4c14dec0df01a8eadab587920753fee67abf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92fbcf119fd704d01bca613b32439360422fe8ea79ba3331eace6d7f52cb5d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7p6zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.830693 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs\") pod \"network-metrics-daemon-6ft2z\" (UID: \"cd9aa596-76c2-468c-b732-d85bc1f16070\") " pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.830790 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q78q2\" (UniqueName: \"kubernetes.io/projected/cd9aa596-76c2-468c-b732-d85bc1f16070-kube-api-access-q78q2\") pod \"network-metrics-daemon-6ft2z\" (UID: \"cd9aa596-76c2-468c-b732-d85bc1f16070\") " pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:32 crc kubenswrapper[4802]: E1125 16:47:32.831116 4802 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 16:47:32 crc kubenswrapper[4802]: E1125 16:47:32.831333 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs podName:cd9aa596-76c2-468c-b732-d85bc1f16070 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:33.331302061 +0000 UTC m=+36.475649247 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs") pod "network-metrics-daemon-6ft2z" (UID: "cd9aa596-76c2-468c-b732-d85bc1f16070") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.842920 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:28Z\\\",\\\"message\\\":\\\"ce\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}\\\\nF1125 16:47:28.617785 6189 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z]\\\\nI1125 16:47:28.617784 6189 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-service-ca-operator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid ==\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.848139 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q78q2\" (UniqueName: \"kubernetes.io/projected/cd9aa596-76c2-468c-b732-d85bc1f16070-kube-api-access-q78q2\") pod \"network-metrics-daemon-6ft2z\" (UID: \"cd9aa596-76c2-468c-b732-d85bc1f16070\") " pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.852041 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.864799 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.874242 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.874341 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.874370 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.874427 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.874441 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:32Z","lastTransitionTime":"2025-11-25T16:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.876833 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.889294 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.899722 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.910320 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.923916 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.937276 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.949085 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.961750 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.974298 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.976900 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.976944 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.976953 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.976970 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.976980 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:32Z","lastTransitionTime":"2025-11-25T16:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.986031 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:32 crc kubenswrapper[4802]: I1125 16:47:32.999642 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.009621 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd9aa596-76c2-468c-b732-d85bc1f16070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6ft2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:33Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.080091 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.080183 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.080198 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.080219 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.080232 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:33Z","lastTransitionTime":"2025-11-25T16:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.182107 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.182156 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.182165 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.182181 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.182191 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:33Z","lastTransitionTime":"2025-11-25T16:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.207030 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.207075 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.207090 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.207108 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.207132 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:33Z","lastTransitionTime":"2025-11-25T16:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:33 crc kubenswrapper[4802]: E1125 16:47:33.219473 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:33Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.222705 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.222744 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.222757 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.222773 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.222784 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:33Z","lastTransitionTime":"2025-11-25T16:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:33 crc kubenswrapper[4802]: E1125 16:47:33.233596 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:33Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.237479 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.237513 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.237527 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.237546 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.237557 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:33Z","lastTransitionTime":"2025-11-25T16:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:33 crc kubenswrapper[4802]: E1125 16:47:33.248044 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:33Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.251142 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.251182 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.251193 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.251208 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.251223 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:33Z","lastTransitionTime":"2025-11-25T16:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:33 crc kubenswrapper[4802]: E1125 16:47:33.262496 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:33Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.266516 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.266553 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.266563 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.266580 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.266589 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:33Z","lastTransitionTime":"2025-11-25T16:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:33 crc kubenswrapper[4802]: E1125 16:47:33.278588 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:33Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:33 crc kubenswrapper[4802]: E1125 16:47:33.278741 4802 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.283774 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.283833 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.283847 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.283866 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.283878 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:33Z","lastTransitionTime":"2025-11-25T16:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.337104 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs\") pod \"network-metrics-daemon-6ft2z\" (UID: \"cd9aa596-76c2-468c-b732-d85bc1f16070\") " pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:33 crc kubenswrapper[4802]: E1125 16:47:33.337286 4802 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 16:47:33 crc kubenswrapper[4802]: E1125 16:47:33.337356 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs podName:cd9aa596-76c2-468c-b732-d85bc1f16070 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:34.337338287 +0000 UTC m=+37.481685473 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs") pod "network-metrics-daemon-6ft2z" (UID: "cd9aa596-76c2-468c-b732-d85bc1f16070") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.386876 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.386924 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.386941 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.386975 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.386992 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:33Z","lastTransitionTime":"2025-11-25T16:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.489945 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.489990 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.490004 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.490029 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.490051 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:33Z","lastTransitionTime":"2025-11-25T16:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.503635 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:33 crc kubenswrapper[4802]: E1125 16:47:33.503756 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.593087 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.593149 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.593157 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.593173 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.593182 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:33Z","lastTransitionTime":"2025-11-25T16:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.695233 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.695278 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.695286 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.695301 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.695310 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:33Z","lastTransitionTime":"2025-11-25T16:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.797937 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.797987 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.798001 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.798018 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.798030 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:33Z","lastTransitionTime":"2025-11-25T16:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.899807 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.899842 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.899852 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.899865 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:33 crc kubenswrapper[4802]: I1125 16:47:33.899873 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:33Z","lastTransitionTime":"2025-11-25T16:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.001698 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.001738 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.001748 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.001763 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.001775 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:34Z","lastTransitionTime":"2025-11-25T16:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.104073 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.104143 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.104158 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.104173 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.104183 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:34Z","lastTransitionTime":"2025-11-25T16:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.206679 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.206724 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.206741 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.206759 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.206771 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:34Z","lastTransitionTime":"2025-11-25T16:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.309700 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.309758 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.309822 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.309843 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.309854 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:34Z","lastTransitionTime":"2025-11-25T16:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.345835 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs\") pod \"network-metrics-daemon-6ft2z\" (UID: \"cd9aa596-76c2-468c-b732-d85bc1f16070\") " pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:34 crc kubenswrapper[4802]: E1125 16:47:34.346051 4802 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 16:47:34 crc kubenswrapper[4802]: E1125 16:47:34.346224 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs podName:cd9aa596-76c2-468c-b732-d85bc1f16070 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:36.346196718 +0000 UTC m=+39.490543904 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs") pod "network-metrics-daemon-6ft2z" (UID: "cd9aa596-76c2-468c-b732-d85bc1f16070") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.412879 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.412948 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.412960 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.412996 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.413011 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:34Z","lastTransitionTime":"2025-11-25T16:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.504065 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.504161 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.504087 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:34 crc kubenswrapper[4802]: E1125 16:47:34.504299 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:47:34 crc kubenswrapper[4802]: E1125 16:47:34.504350 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:34 crc kubenswrapper[4802]: E1125 16:47:34.504410 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.515403 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.515448 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.515463 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.515482 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.515496 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:34Z","lastTransitionTime":"2025-11-25T16:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.618063 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.618563 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.618636 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.618738 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.618826 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:34Z","lastTransitionTime":"2025-11-25T16:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.721261 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.721349 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.721363 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.721388 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.721405 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:34Z","lastTransitionTime":"2025-11-25T16:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.825254 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.825308 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.825321 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.825342 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.825356 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:34Z","lastTransitionTime":"2025-11-25T16:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.928268 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.928346 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.928365 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.928391 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:34 crc kubenswrapper[4802]: I1125 16:47:34.928408 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:34Z","lastTransitionTime":"2025-11-25T16:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.031382 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.031462 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.031476 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.031499 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.031513 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:35Z","lastTransitionTime":"2025-11-25T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.134440 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.134505 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.134518 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.134540 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.134552 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:35Z","lastTransitionTime":"2025-11-25T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.237588 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.237654 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.237671 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.237687 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.237697 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:35Z","lastTransitionTime":"2025-11-25T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.340385 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.340427 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.340442 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.340460 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.340470 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:35Z","lastTransitionTime":"2025-11-25T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.443065 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.443117 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.443150 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.443169 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.443180 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:35Z","lastTransitionTime":"2025-11-25T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.504324 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:35 crc kubenswrapper[4802]: E1125 16:47:35.504507 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.546212 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.546300 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.546347 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.546389 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.546419 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:35Z","lastTransitionTime":"2025-11-25T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.649308 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.649370 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.649380 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.649403 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.649415 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:35Z","lastTransitionTime":"2025-11-25T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.751758 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.751801 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.751809 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.751823 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.751833 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:35Z","lastTransitionTime":"2025-11-25T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.854903 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.855518 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.855566 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.855604 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.855623 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:35Z","lastTransitionTime":"2025-11-25T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.959469 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.959519 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.959530 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.959544 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:35 crc kubenswrapper[4802]: I1125 16:47:35.959554 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:35Z","lastTransitionTime":"2025-11-25T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.063259 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.063302 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.063312 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.063330 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.063349 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:36Z","lastTransitionTime":"2025-11-25T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.166026 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.166081 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.166095 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.166116 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.166153 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:36Z","lastTransitionTime":"2025-11-25T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.268286 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.268329 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.268340 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.268359 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.268375 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:36Z","lastTransitionTime":"2025-11-25T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.367676 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs\") pod \"network-metrics-daemon-6ft2z\" (UID: \"cd9aa596-76c2-468c-b732-d85bc1f16070\") " pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:36 crc kubenswrapper[4802]: E1125 16:47:36.367930 4802 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 16:47:36 crc kubenswrapper[4802]: E1125 16:47:36.368041 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs podName:cd9aa596-76c2-468c-b732-d85bc1f16070 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:40.368021224 +0000 UTC m=+43.512368410 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs") pod "network-metrics-daemon-6ft2z" (UID: "cd9aa596-76c2-468c-b732-d85bc1f16070") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.371529 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.371589 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.371604 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.371685 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.371708 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:36Z","lastTransitionTime":"2025-11-25T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.474283 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.474335 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.474350 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.474371 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.474384 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:36Z","lastTransitionTime":"2025-11-25T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.503933 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.504044 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.504055 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:36 crc kubenswrapper[4802]: E1125 16:47:36.504336 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:36 crc kubenswrapper[4802]: E1125 16:47:36.504564 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:36 crc kubenswrapper[4802]: E1125 16:47:36.504735 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.579000 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.579062 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.579083 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.579112 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.579156 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:36Z","lastTransitionTime":"2025-11-25T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.682304 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.682368 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.682380 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.682402 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.682415 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:36Z","lastTransitionTime":"2025-11-25T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.786087 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.786223 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.786245 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.786280 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.786302 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:36Z","lastTransitionTime":"2025-11-25T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.889155 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.889200 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.889210 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.889246 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.889256 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:36Z","lastTransitionTime":"2025-11-25T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.992987 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.993043 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.993067 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.993095 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:36 crc kubenswrapper[4802]: I1125 16:47:36.993113 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:36Z","lastTransitionTime":"2025-11-25T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.095777 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.095823 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.095833 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.095848 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.095862 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:37Z","lastTransitionTime":"2025-11-25T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.198512 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.198577 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.198589 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.198618 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.198634 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:37Z","lastTransitionTime":"2025-11-25T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.301270 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.301347 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.301359 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.301381 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.301396 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:37Z","lastTransitionTime":"2025-11-25T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.404629 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.404678 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.404696 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.404719 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.404735 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:37Z","lastTransitionTime":"2025-11-25T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.504008 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:37 crc kubenswrapper[4802]: E1125 16:47:37.504222 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.507361 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.507425 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.507444 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.507471 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.507493 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:37Z","lastTransitionTime":"2025-11-25T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.520617 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:37Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.532658 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:37Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.548891 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:37Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.565035 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:37Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.583038 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:37Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.600182 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:37Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.609321 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.609373 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.609382 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.609398 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.609408 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:37Z","lastTransitionTime":"2025-11-25T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.612396 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:37Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.624773 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:37Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.635776 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd9aa596-76c2-468c-b732-d85bc1f16070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6ft2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:37Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.648350 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feaf869b-f88a-4874-8174-b689fcf6f40a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d672b4cc7df9d1855c8e882c5f4c14dec0df01a8eadab587920753fee67abf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92fbcf119fd704d01bca613b32439360422fe8ea79ba3331eace6d7f52cb5d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7p6zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:37Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.667785 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:28Z\\\",\\\"message\\\":\\\"ce\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}\\\\nF1125 16:47:28.617785 6189 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z]\\\\nI1125 16:47:28.617784 6189 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-service-ca-operator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid ==\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:37Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.680179 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:37Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.693416 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:37Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.707080 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:37Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.711398 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.711446 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.711458 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.711477 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.711491 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:37Z","lastTransitionTime":"2025-11-25T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.719797 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:37Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.729883 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:37Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.813865 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.813900 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.813911 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.813942 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.813952 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:37Z","lastTransitionTime":"2025-11-25T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.916438 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.916496 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.916507 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.916539 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:37 crc kubenswrapper[4802]: I1125 16:47:37.916554 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:37Z","lastTransitionTime":"2025-11-25T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.018898 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.018938 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.018948 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.018961 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.018974 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:38Z","lastTransitionTime":"2025-11-25T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.121406 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.121483 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.121493 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.121510 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.121522 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:38Z","lastTransitionTime":"2025-11-25T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.223946 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.223979 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.223989 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.224003 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.224012 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:38Z","lastTransitionTime":"2025-11-25T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.326791 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.326862 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.326882 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.326897 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.326910 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:38Z","lastTransitionTime":"2025-11-25T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.429224 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.429263 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.429273 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.429290 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.429299 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:38Z","lastTransitionTime":"2025-11-25T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.503478 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.503503 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.503503 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:38 crc kubenswrapper[4802]: E1125 16:47:38.503608 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:38 crc kubenswrapper[4802]: E1125 16:47:38.503707 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:38 crc kubenswrapper[4802]: E1125 16:47:38.503788 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.531202 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.531246 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.531274 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.531290 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.531299 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:38Z","lastTransitionTime":"2025-11-25T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.633435 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.633472 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.633483 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.633498 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.633509 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:38Z","lastTransitionTime":"2025-11-25T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.735232 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.735319 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.735336 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.735361 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.735372 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:38Z","lastTransitionTime":"2025-11-25T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.837214 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.837282 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.837301 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.837326 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.837344 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:38Z","lastTransitionTime":"2025-11-25T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.939256 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.939299 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.939308 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.939321 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:38 crc kubenswrapper[4802]: I1125 16:47:38.939333 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:38Z","lastTransitionTime":"2025-11-25T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.041609 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.041644 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.041655 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.041670 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.041681 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:39Z","lastTransitionTime":"2025-11-25T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.144529 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.144587 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.144601 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.144621 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.144636 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:39Z","lastTransitionTime":"2025-11-25T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.246702 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.246733 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.246741 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.246753 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.246763 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:39Z","lastTransitionTime":"2025-11-25T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.349262 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.349325 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.349338 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.349355 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.349367 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:39Z","lastTransitionTime":"2025-11-25T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.451978 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.452296 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.452307 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.452320 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.452329 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:39Z","lastTransitionTime":"2025-11-25T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.503887 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:39 crc kubenswrapper[4802]: E1125 16:47:39.504017 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.554945 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.554982 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.554993 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.555008 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.555018 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:39Z","lastTransitionTime":"2025-11-25T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.657381 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.657430 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.657442 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.657460 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.657471 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:39Z","lastTransitionTime":"2025-11-25T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.760222 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.760284 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.760300 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.760321 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.760335 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:39Z","lastTransitionTime":"2025-11-25T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.862663 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.862711 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.862724 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.862769 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.862784 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:39Z","lastTransitionTime":"2025-11-25T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.965005 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.965049 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.965059 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.965074 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:39 crc kubenswrapper[4802]: I1125 16:47:39.965084 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:39Z","lastTransitionTime":"2025-11-25T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.067384 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.067426 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.067438 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.067455 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.067467 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:40Z","lastTransitionTime":"2025-11-25T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.169430 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.169486 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.169496 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.169511 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.169521 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:40Z","lastTransitionTime":"2025-11-25T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.271274 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.271307 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.271318 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.271333 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.271344 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:40Z","lastTransitionTime":"2025-11-25T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.373304 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.373367 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.373391 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.373416 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.373435 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:40Z","lastTransitionTime":"2025-11-25T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.407960 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs\") pod \"network-metrics-daemon-6ft2z\" (UID: \"cd9aa596-76c2-468c-b732-d85bc1f16070\") " pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:40 crc kubenswrapper[4802]: E1125 16:47:40.408162 4802 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 16:47:40 crc kubenswrapper[4802]: E1125 16:47:40.408264 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs podName:cd9aa596-76c2-468c-b732-d85bc1f16070 nodeName:}" failed. No retries permitted until 2025-11-25 16:47:48.408240429 +0000 UTC m=+51.552587625 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs") pod "network-metrics-daemon-6ft2z" (UID: "cd9aa596-76c2-468c-b732-d85bc1f16070") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.475651 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.475682 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.475690 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.475720 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.475731 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:40Z","lastTransitionTime":"2025-11-25T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.504057 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.504057 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:40 crc kubenswrapper[4802]: E1125 16:47:40.504175 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.504233 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:40 crc kubenswrapper[4802]: E1125 16:47:40.504297 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:40 crc kubenswrapper[4802]: E1125 16:47:40.504438 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.578158 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.578197 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.578224 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.578241 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.578253 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:40Z","lastTransitionTime":"2025-11-25T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.681264 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.681310 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.681321 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.681338 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.681353 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:40Z","lastTransitionTime":"2025-11-25T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.783345 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.783386 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.783397 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.783413 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.783424 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:40Z","lastTransitionTime":"2025-11-25T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.886244 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.886311 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.886330 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.886356 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.886374 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:40Z","lastTransitionTime":"2025-11-25T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.989490 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.989548 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.989559 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.989578 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:40 crc kubenswrapper[4802]: I1125 16:47:40.989589 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:40Z","lastTransitionTime":"2025-11-25T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.092388 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.092440 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.092452 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.092469 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.092481 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:41Z","lastTransitionTime":"2025-11-25T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.195189 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.195250 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.195262 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.195279 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.195290 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:41Z","lastTransitionTime":"2025-11-25T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.297621 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.297675 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.297689 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.297708 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.297722 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:41Z","lastTransitionTime":"2025-11-25T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.400379 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.400431 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.400447 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.400470 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.400482 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:41Z","lastTransitionTime":"2025-11-25T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.503471 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.503517 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.503527 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.503541 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.503549 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:41Z","lastTransitionTime":"2025-11-25T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.505902 4802 scope.go:117] "RemoveContainer" containerID="eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.506769 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:41 crc kubenswrapper[4802]: E1125 16:47:41.507336 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.605899 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.605943 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.605956 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.605993 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.606005 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:41Z","lastTransitionTime":"2025-11-25T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.710216 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.710252 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.710262 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.710278 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.710290 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:41Z","lastTransitionTime":"2025-11-25T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.728150 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovnkube-controller/1.log" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.731515 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerStarted","Data":"5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977"} Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.732110 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.754942 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:28Z\\\",\\\"message\\\":\\\"ce\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}\\\\nF1125 16:47:28.617785 6189 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z]\\\\nI1125 16:47:28.617784 6189 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-service-ca-operator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid ==\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.768302 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.785973 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.800869 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.812340 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.812397 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.812409 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.812435 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.812451 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:41Z","lastTransitionTime":"2025-11-25T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.817683 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.837822 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.853955 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.872451 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.889861 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.909018 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.914694 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.914754 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.914771 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.914798 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.914814 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:41Z","lastTransitionTime":"2025-11-25T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.927779 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.941045 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.953899 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.968284 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.980620 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd9aa596-76c2-468c-b732-d85bc1f16070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6ft2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:41 crc kubenswrapper[4802]: I1125 16:47:41.995267 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feaf869b-f88a-4874-8174-b689fcf6f40a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d672b4cc7df9d1855c8e882c5f4c14dec0df01a8eadab587920753fee67abf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92fbcf119fd704d01bca613b32439360422fe8ea79ba3331eace6d7f52cb5d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7p6zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.017284 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.017346 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.017357 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.017382 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.017400 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:42Z","lastTransitionTime":"2025-11-25T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.119546 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.119607 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.119619 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.119643 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.119657 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:42Z","lastTransitionTime":"2025-11-25T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.221820 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.221871 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.221882 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.221905 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.221918 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:42Z","lastTransitionTime":"2025-11-25T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.323860 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.323902 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.323912 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.323925 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.323934 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:42Z","lastTransitionTime":"2025-11-25T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.426751 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.426799 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.426813 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.426831 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.426843 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:42Z","lastTransitionTime":"2025-11-25T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.504064 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.504107 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.504139 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:42 crc kubenswrapper[4802]: E1125 16:47:42.504222 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:47:42 crc kubenswrapper[4802]: E1125 16:47:42.504335 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:42 crc kubenswrapper[4802]: E1125 16:47:42.504438 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.529489 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.529546 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.529562 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.529586 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.529603 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:42Z","lastTransitionTime":"2025-11-25T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.632768 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.632816 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.632826 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.632840 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.632851 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:42Z","lastTransitionTime":"2025-11-25T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.734661 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.735764 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.735808 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.735834 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.735858 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:42Z","lastTransitionTime":"2025-11-25T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.738241 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovnkube-controller/2.log" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.739786 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovnkube-controller/1.log" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.743147 4802 generic.go:334] "Generic (PLEG): container finished" podID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerID="5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977" exitCode=1 Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.743199 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerDied","Data":"5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977"} Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.743273 4802 scope.go:117] "RemoveContainer" containerID="eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.743996 4802 scope.go:117] "RemoveContainer" containerID="5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977" Nov 25 16:47:42 crc kubenswrapper[4802]: E1125 16:47:42.744223 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.759185 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.773998 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.787799 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.800244 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.811756 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.824544 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.836800 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.838718 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.838756 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.838767 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.838783 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.838794 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:42Z","lastTransitionTime":"2025-11-25T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.853323 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.870278 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.884526 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.899377 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.911717 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.922152 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd9aa596-76c2-468c-b732-d85bc1f16070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6ft2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.933473 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feaf869b-f88a-4874-8174-b689fcf6f40a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d672b4cc7df9d1855c8e882c5f4c14dec0df01a8eadab587920753fee67abf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92fbcf119fd704d01bca613b32439360422fe8ea79ba3331eace6d7f52cb5d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7p6zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.940646 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.940677 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.940687 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.940711 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.940721 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:42Z","lastTransitionTime":"2025-11-25T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.956863 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb6d42a574093bceacfae9d03a7430bed9f0a3ae76062063b5ad50731dafa7fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:28Z\\\",\\\"message\\\":\\\"ce\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}\\\\nF1125 16:47:28.617785 6189 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:28Z is after 2025-08-24T17:21:41Z]\\\\nI1125 16:47:28.617784 6189 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-service-ca-operator/metrics]} name:Service_openshift-service-ca-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid ==\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:42Z\\\",\\\"message\\\":\\\"4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1125 16:47:42.255841 6417 lb_config.go:1031] Cluster endpoints for openshift-console-operator/metrics for network=default are: map[]\\\\nF1125 16:47:42.255847 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z]\\\\nI1125 16:47:42.255857 6417 services_controller.go:443] Built service openshift-console-operator/metrics LB cluster-wide con\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:42 crc kubenswrapper[4802]: I1125 16:47:42.970754 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.042996 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.043040 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.043057 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.043075 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.043086 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:43Z","lastTransitionTime":"2025-11-25T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.146445 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.146507 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.146520 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.146541 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.146555 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:43Z","lastTransitionTime":"2025-11-25T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.249336 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.249377 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.249390 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.249409 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.249421 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:43Z","lastTransitionTime":"2025-11-25T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.352256 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.352299 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.352310 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.352325 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.352337 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:43Z","lastTransitionTime":"2025-11-25T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.454689 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.454937 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.455024 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.455093 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.455252 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:43Z","lastTransitionTime":"2025-11-25T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.461255 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.461374 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.461452 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.461522 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.461579 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:43Z","lastTransitionTime":"2025-11-25T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:43 crc kubenswrapper[4802]: E1125 16:47:43.474464 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.478373 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.478568 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.478644 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.478725 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.478833 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:43Z","lastTransitionTime":"2025-11-25T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:43 crc kubenswrapper[4802]: E1125 16:47:43.494192 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.499722 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.499772 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.499789 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.499812 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.499828 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:43Z","lastTransitionTime":"2025-11-25T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.505274 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:43 crc kubenswrapper[4802]: E1125 16:47:43.505404 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:43 crc kubenswrapper[4802]: E1125 16:47:43.516194 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.519875 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.519914 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.519926 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.519943 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.519953 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:43Z","lastTransitionTime":"2025-11-25T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:43 crc kubenswrapper[4802]: E1125 16:47:43.535077 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.538670 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.538713 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.538726 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.538746 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.538759 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:43Z","lastTransitionTime":"2025-11-25T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:43 crc kubenswrapper[4802]: E1125 16:47:43.553368 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: E1125 16:47:43.553478 4802 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.558083 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.558153 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.558170 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.558191 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.558203 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:43Z","lastTransitionTime":"2025-11-25T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.660153 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.660202 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.660217 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.660236 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.660249 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:43Z","lastTransitionTime":"2025-11-25T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.748264 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovnkube-controller/2.log" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.751961 4802 scope.go:117] "RemoveContainer" containerID="5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977" Nov 25 16:47:43 crc kubenswrapper[4802]: E1125 16:47:43.752217 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.762741 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.762787 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.762802 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.762822 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.762837 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:43Z","lastTransitionTime":"2025-11-25T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.765387 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.778543 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.789576 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.799727 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.810593 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.823268 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.833720 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.843964 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.854847 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd9aa596-76c2-468c-b732-d85bc1f16070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6ft2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.864500 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.864549 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.864562 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.864578 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.864590 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:43Z","lastTransitionTime":"2025-11-25T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.871964 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feaf869b-f88a-4874-8174-b689fcf6f40a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d672b4cc7df9d1855c8e882c5f4c14dec0df01a8eadab587920753fee67abf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92fbcf119fd704d01bca613b32439360422fe8ea79ba3331eace6d7f52cb5d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7p6zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.890993 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:42Z\\\",\\\"message\\\":\\\"4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1125 16:47:42.255841 6417 lb_config.go:1031] Cluster endpoints for openshift-console-operator/metrics for network=default are: map[]\\\\nF1125 16:47:42.255847 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z]\\\\nI1125 16:47:42.255857 6417 services_controller.go:443] Built service openshift-console-operator/metrics LB cluster-wide con\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.899807 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.910182 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.924614 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.937239 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.948673 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:43Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.967310 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.967352 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.967364 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.967381 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:43 crc kubenswrapper[4802]: I1125 16:47:43.967395 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:43Z","lastTransitionTime":"2025-11-25T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.069737 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.069777 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.069785 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.069803 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.069814 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:44Z","lastTransitionTime":"2025-11-25T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.172493 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.172542 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.172552 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.172569 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.172579 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:44Z","lastTransitionTime":"2025-11-25T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.275544 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.275593 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.275603 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.275621 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.275633 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:44Z","lastTransitionTime":"2025-11-25T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.378564 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.378599 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.378610 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.378626 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.378637 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:44Z","lastTransitionTime":"2025-11-25T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.480911 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.480971 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.480982 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.480999 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.481010 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:44Z","lastTransitionTime":"2025-11-25T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.503566 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:44 crc kubenswrapper[4802]: E1125 16:47:44.503850 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.504770 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:44 crc kubenswrapper[4802]: E1125 16:47:44.504869 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.504942 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:44 crc kubenswrapper[4802]: E1125 16:47:44.505080 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.583593 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.583638 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.583649 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.583666 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.583677 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:44Z","lastTransitionTime":"2025-11-25T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.685589 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.685665 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.685677 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.685696 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.685709 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:44Z","lastTransitionTime":"2025-11-25T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.787764 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.787807 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.787818 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.787832 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.787843 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:44Z","lastTransitionTime":"2025-11-25T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.890287 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.890330 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.890340 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.890354 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.890363 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:44Z","lastTransitionTime":"2025-11-25T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.991990 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.992057 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.992078 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.992171 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:44 crc kubenswrapper[4802]: I1125 16:47:44.992199 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:44Z","lastTransitionTime":"2025-11-25T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.095424 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.095527 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.095552 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.095585 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.095621 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:45Z","lastTransitionTime":"2025-11-25T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.197811 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.197880 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.197895 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.197921 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.197948 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:45Z","lastTransitionTime":"2025-11-25T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.300055 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.300102 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.300113 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.300156 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.300168 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:45Z","lastTransitionTime":"2025-11-25T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.402734 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.402799 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.402816 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.402832 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.402844 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:45Z","lastTransitionTime":"2025-11-25T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.504053 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:45 crc kubenswrapper[4802]: E1125 16:47:45.504275 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.505426 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.505454 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.505464 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.505477 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.505488 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:45Z","lastTransitionTime":"2025-11-25T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.607764 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.607810 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.607822 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.607839 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.607851 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:45Z","lastTransitionTime":"2025-11-25T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.710478 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.710529 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.710538 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.710551 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.710560 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:45Z","lastTransitionTime":"2025-11-25T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.798371 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.804833 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.812247 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:45Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.812499 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.812533 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.812543 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.812562 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.812574 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:45Z","lastTransitionTime":"2025-11-25T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.824530 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:45Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.835810 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:45Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.848319 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:45Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.860434 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:45Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.871037 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:45Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.886051 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:45Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.898349 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:45Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.908070 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd9aa596-76c2-468c-b732-d85bc1f16070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6ft2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:45Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.914315 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.914371 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.914381 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.914399 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.914412 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:45Z","lastTransitionTime":"2025-11-25T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.922906 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feaf869b-f88a-4874-8174-b689fcf6f40a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d672b4cc7df9d1855c8e882c5f4c14dec0df01a8eadab587920753fee67abf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92fbcf119fd704d01bca613b32439360422fe8ea79ba3331eace6d7f52cb5d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7p6zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:45Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.947408 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:42Z\\\",\\\"message\\\":\\\"4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1125 16:47:42.255841 6417 lb_config.go:1031] Cluster endpoints for openshift-console-operator/metrics for network=default are: map[]\\\\nF1125 16:47:42.255847 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z]\\\\nI1125 16:47:42.255857 6417 services_controller.go:443] Built service openshift-console-operator/metrics LB cluster-wide con\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:45Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.959531 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:45Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.972223 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:45Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.983822 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:45Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:45 crc kubenswrapper[4802]: I1125 16:47:45.992024 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:45Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.003303 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:46Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.017422 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.017454 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.017466 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.017480 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.017491 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:46Z","lastTransitionTime":"2025-11-25T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.119696 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.119746 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.119758 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.119775 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.119785 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:46Z","lastTransitionTime":"2025-11-25T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.223299 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.223392 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.223408 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.223430 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.223444 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:46Z","lastTransitionTime":"2025-11-25T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.326339 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.326399 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.326413 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.326436 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.326452 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:46Z","lastTransitionTime":"2025-11-25T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.429719 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.429776 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.429798 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.429820 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.429837 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:46Z","lastTransitionTime":"2025-11-25T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.503776 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.503849 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:46 crc kubenswrapper[4802]: E1125 16:47:46.503940 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.503950 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:46 crc kubenswrapper[4802]: E1125 16:47:46.504035 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:47:46 crc kubenswrapper[4802]: E1125 16:47:46.504107 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.532324 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.532370 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.532381 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.532396 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.532408 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:46Z","lastTransitionTime":"2025-11-25T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.635193 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.635281 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.635300 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.635326 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.635346 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:46Z","lastTransitionTime":"2025-11-25T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.737793 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.737833 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.737842 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.737855 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.737865 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:46Z","lastTransitionTime":"2025-11-25T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.840495 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.840548 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.840567 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.840591 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.840607 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:46Z","lastTransitionTime":"2025-11-25T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.943250 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.943321 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.943332 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.943351 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:46 crc kubenswrapper[4802]: I1125 16:47:46.943366 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:46Z","lastTransitionTime":"2025-11-25T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.045424 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.045462 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.045474 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.045491 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.045501 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:47Z","lastTransitionTime":"2025-11-25T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.147488 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.147539 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.147550 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.147572 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.147585 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:47Z","lastTransitionTime":"2025-11-25T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.249826 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.250115 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.250239 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.250340 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.250426 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:47Z","lastTransitionTime":"2025-11-25T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.353735 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.354180 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.354204 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.354231 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.354249 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:47Z","lastTransitionTime":"2025-11-25T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.456677 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.456723 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.456735 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.456755 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.456768 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:47Z","lastTransitionTime":"2025-11-25T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.504098 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:47 crc kubenswrapper[4802]: E1125 16:47:47.504260 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.519281 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.530028 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.541563 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.553994 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.559081 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.559166 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.559186 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.559210 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.559268 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:47Z","lastTransitionTime":"2025-11-25T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.569480 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.584975 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.600198 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.610538 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.621626 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.651092 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.662900 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.662965 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.662978 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.662997 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.663009 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:47Z","lastTransitionTime":"2025-11-25T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.679362 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.693923 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.704801 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd9aa596-76c2-468c-b732-d85bc1f16070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6ft2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.716545 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fd65faa-292f-433b-83de-043c66cf484d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://060bdec5b0ff95b1ac4dcdb60f4bbb811ddeff0c93e74e373640bf65cb1d5ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb75f902fead2f9945afbecfdfd7b7fd35c5eaa4a05059a2038fb08e07875ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4bee24ceffdebad39410dfdca5f8a83c8664f6b4c18cc2495f47f47162d116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.727769 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feaf869b-f88a-4874-8174-b689fcf6f40a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d672b4cc7df9d1855c8e882c5f4c14dec0df01a8eadab587920753fee67abf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92fbcf119fd704d01bca613b32439360422fe8ea79ba3331eace6d7f52cb5d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7p6zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.750492 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:42Z\\\",\\\"message\\\":\\\"4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1125 16:47:42.255841 6417 lb_config.go:1031] Cluster endpoints for openshift-console-operator/metrics for network=default are: map[]\\\\nF1125 16:47:42.255847 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z]\\\\nI1125 16:47:42.255857 6417 services_controller.go:443] Built service openshift-console-operator/metrics LB cluster-wide con\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.761248 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.765319 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.765355 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.765367 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.765384 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.765396 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:47Z","lastTransitionTime":"2025-11-25T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.868543 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.868941 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.869107 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.869282 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.869446 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:47Z","lastTransitionTime":"2025-11-25T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.972166 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.972429 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.972508 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.972615 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:47 crc kubenswrapper[4802]: I1125 16:47:47.972710 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:47Z","lastTransitionTime":"2025-11-25T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.075100 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.075186 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.075198 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.075238 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.075254 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:48Z","lastTransitionTime":"2025-11-25T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.177754 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.177798 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.177808 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.177823 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.177834 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:48Z","lastTransitionTime":"2025-11-25T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.280950 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.281026 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.281045 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.281071 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.281089 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:48Z","lastTransitionTime":"2025-11-25T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.383428 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.383767 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.383884 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.383980 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.384065 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:48Z","lastTransitionTime":"2025-11-25T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.486725 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.486776 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.486790 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.486807 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.486818 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:48Z","lastTransitionTime":"2025-11-25T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.497341 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.497425 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.497465 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.497484 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.497502 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs\") pod \"network-metrics-daemon-6ft2z\" (UID: \"cd9aa596-76c2-468c-b732-d85bc1f16070\") " pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:48 crc kubenswrapper[4802]: E1125 16:47:48.497587 4802 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 16:47:48 crc kubenswrapper[4802]: E1125 16:47:48.497625 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs podName:cd9aa596-76c2-468c-b732-d85bc1f16070 nodeName:}" failed. No retries permitted until 2025-11-25 16:48:04.497612462 +0000 UTC m=+67.641959648 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs") pod "network-metrics-daemon-6ft2z" (UID: "cd9aa596-76c2-468c-b732-d85bc1f16070") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 16:47:48 crc kubenswrapper[4802]: E1125 16:47:48.497801 4802 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 16:47:48 crc kubenswrapper[4802]: E1125 16:47:48.497843 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 16:48:20.497834748 +0000 UTC m=+83.642181934 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 16:47:48 crc kubenswrapper[4802]: E1125 16:47:48.497891 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:48:20.497882229 +0000 UTC m=+83.642229405 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:47:48 crc kubenswrapper[4802]: E1125 16:47:48.497949 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 16:47:48 crc kubenswrapper[4802]: E1125 16:47:48.497971 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 16:47:48 crc kubenswrapper[4802]: E1125 16:47:48.497982 4802 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:48 crc kubenswrapper[4802]: E1125 16:47:48.498009 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 16:48:20.498002472 +0000 UTC m=+83.642349648 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:48 crc kubenswrapper[4802]: E1125 16:47:48.498117 4802 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 16:47:48 crc kubenswrapper[4802]: E1125 16:47:48.498269 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 16:48:20.498251618 +0000 UTC m=+83.642598804 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.504206 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:48 crc kubenswrapper[4802]: E1125 16:47:48.504319 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.504223 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:48 crc kubenswrapper[4802]: E1125 16:47:48.504397 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.504206 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:48 crc kubenswrapper[4802]: E1125 16:47:48.504442 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.589374 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.589417 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.589427 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.589442 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.589455 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:48Z","lastTransitionTime":"2025-11-25T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.597951 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:48 crc kubenswrapper[4802]: E1125 16:47:48.598091 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 16:47:48 crc kubenswrapper[4802]: E1125 16:47:48.598112 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 16:47:48 crc kubenswrapper[4802]: E1125 16:47:48.598143 4802 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:48 crc kubenswrapper[4802]: E1125 16:47:48.598190 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 16:48:20.598177364 +0000 UTC m=+83.742524550 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.692280 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.692317 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.692326 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.692339 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.692350 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:48Z","lastTransitionTime":"2025-11-25T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.794915 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.794963 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.794976 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.794992 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.795004 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:48Z","lastTransitionTime":"2025-11-25T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.897629 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.897677 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.897691 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.897712 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:48 crc kubenswrapper[4802]: I1125 16:47:48.897726 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:48Z","lastTransitionTime":"2025-11-25T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.000186 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.000212 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.000242 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.000256 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.000266 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:49Z","lastTransitionTime":"2025-11-25T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.102849 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.102886 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.102895 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.102907 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.102916 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:49Z","lastTransitionTime":"2025-11-25T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.204580 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.204613 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.204623 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.204635 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.204644 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:49Z","lastTransitionTime":"2025-11-25T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.306408 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.306462 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.306480 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.306502 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.306556 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:49Z","lastTransitionTime":"2025-11-25T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.408691 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.408751 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.408768 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.408790 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.408805 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:49Z","lastTransitionTime":"2025-11-25T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.504625 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:49 crc kubenswrapper[4802]: E1125 16:47:49.504860 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.510870 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.510898 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.510906 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.510920 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.510935 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:49Z","lastTransitionTime":"2025-11-25T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.614072 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.614117 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.614141 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.614155 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.614163 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:49Z","lastTransitionTime":"2025-11-25T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.716633 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.716702 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.716710 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.716726 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.716736 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:49Z","lastTransitionTime":"2025-11-25T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.818879 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.818920 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.818932 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.818949 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.818963 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:49Z","lastTransitionTime":"2025-11-25T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.921841 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.921936 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.921948 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.921960 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:49 crc kubenswrapper[4802]: I1125 16:47:49.921969 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:49Z","lastTransitionTime":"2025-11-25T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.023876 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.023917 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.023929 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.023945 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.023957 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:50Z","lastTransitionTime":"2025-11-25T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.126384 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.126452 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.126470 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.126497 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.126514 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:50Z","lastTransitionTime":"2025-11-25T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.228812 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.228892 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.228921 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.228953 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.228975 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:50Z","lastTransitionTime":"2025-11-25T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.331550 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.331623 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.331636 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.331653 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.331665 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:50Z","lastTransitionTime":"2025-11-25T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.434581 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.434646 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.434663 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.434686 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.434703 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:50Z","lastTransitionTime":"2025-11-25T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.503907 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.503981 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.503924 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:50 crc kubenswrapper[4802]: E1125 16:47:50.504093 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:50 crc kubenswrapper[4802]: E1125 16:47:50.504240 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:50 crc kubenswrapper[4802]: E1125 16:47:50.504390 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.537668 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.537724 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.537736 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.537756 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.537772 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:50Z","lastTransitionTime":"2025-11-25T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.640326 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.640377 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.640388 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.640402 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.640411 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:50Z","lastTransitionTime":"2025-11-25T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.742773 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.742807 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.742817 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.742830 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.742839 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:50Z","lastTransitionTime":"2025-11-25T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.845317 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.845354 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.845363 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.845377 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.845385 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:50Z","lastTransitionTime":"2025-11-25T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.947399 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.947437 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.947448 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.947463 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:50 crc kubenswrapper[4802]: I1125 16:47:50.947474 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:50Z","lastTransitionTime":"2025-11-25T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.049576 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.049624 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.049636 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.049655 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.049667 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:51Z","lastTransitionTime":"2025-11-25T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.153036 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.153082 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.153096 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.153115 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.153152 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:51Z","lastTransitionTime":"2025-11-25T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.255234 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.255259 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.255267 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.255280 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.255288 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:51Z","lastTransitionTime":"2025-11-25T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.357227 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.357280 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.357291 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.357309 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.357322 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:51Z","lastTransitionTime":"2025-11-25T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.459469 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.459508 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.459519 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.459536 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.459546 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:51Z","lastTransitionTime":"2025-11-25T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.504413 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:51 crc kubenswrapper[4802]: E1125 16:47:51.504548 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.561913 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.561940 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.561948 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.561961 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.561970 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:51Z","lastTransitionTime":"2025-11-25T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.664016 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.664062 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.664074 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.664092 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.664111 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:51Z","lastTransitionTime":"2025-11-25T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.766148 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.766182 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.766192 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.766205 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.766214 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:51Z","lastTransitionTime":"2025-11-25T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.867856 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.867918 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.867937 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.867958 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.867973 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:51Z","lastTransitionTime":"2025-11-25T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.969898 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.969930 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.969940 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.969952 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:51 crc kubenswrapper[4802]: I1125 16:47:51.969961 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:51Z","lastTransitionTime":"2025-11-25T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.072037 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.072079 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.072087 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.072102 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.072112 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:52Z","lastTransitionTime":"2025-11-25T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.174259 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.174296 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.174308 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.174324 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.174334 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:52Z","lastTransitionTime":"2025-11-25T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.276556 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.276595 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.276606 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.276621 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.276632 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:52Z","lastTransitionTime":"2025-11-25T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.378612 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.378656 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.378664 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.378680 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.378691 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:52Z","lastTransitionTime":"2025-11-25T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.481154 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.481192 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.481201 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.481216 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.481226 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:52Z","lastTransitionTime":"2025-11-25T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.503749 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.503763 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.503909 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:52 crc kubenswrapper[4802]: E1125 16:47:52.504058 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:52 crc kubenswrapper[4802]: E1125 16:47:52.504190 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:52 crc kubenswrapper[4802]: E1125 16:47:52.504370 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.583025 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.583067 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.583075 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.583088 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.583097 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:52Z","lastTransitionTime":"2025-11-25T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.685300 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.685356 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.685364 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.685377 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.685385 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:52Z","lastTransitionTime":"2025-11-25T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.787943 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.787993 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.788006 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.788023 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.788034 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:52Z","lastTransitionTime":"2025-11-25T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.891098 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.891205 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.891226 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.891249 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.891265 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:52Z","lastTransitionTime":"2025-11-25T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.994445 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.994520 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.994543 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.994571 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:52 crc kubenswrapper[4802]: I1125 16:47:52.994592 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:52Z","lastTransitionTime":"2025-11-25T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.097200 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.097229 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.097237 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.097250 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.097261 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:53Z","lastTransitionTime":"2025-11-25T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.200663 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.200714 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.200733 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.200756 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.200773 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:53Z","lastTransitionTime":"2025-11-25T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.303729 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.303807 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.303828 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.303856 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.303873 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:53Z","lastTransitionTime":"2025-11-25T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.406834 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.406888 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.406901 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.406917 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.406927 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:53Z","lastTransitionTime":"2025-11-25T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.504353 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:53 crc kubenswrapper[4802]: E1125 16:47:53.504588 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.509547 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.509586 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.509597 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.509617 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.509628 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:53Z","lastTransitionTime":"2025-11-25T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.561402 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.561435 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.561451 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.561468 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.561482 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:53Z","lastTransitionTime":"2025-11-25T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:53 crc kubenswrapper[4802]: E1125 16:47:53.577935 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:53Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.582442 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.582488 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.582501 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.582518 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.582530 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:53Z","lastTransitionTime":"2025-11-25T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:53 crc kubenswrapper[4802]: E1125 16:47:53.603234 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:53Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.607624 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.607723 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.607744 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.607768 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.607785 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:53Z","lastTransitionTime":"2025-11-25T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:53 crc kubenswrapper[4802]: E1125 16:47:53.622500 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:53Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.626642 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.626703 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.626722 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.626750 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.626821 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:53Z","lastTransitionTime":"2025-11-25T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:53 crc kubenswrapper[4802]: E1125 16:47:53.640926 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:53Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.643948 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.644016 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.644032 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.644048 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.644061 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:53Z","lastTransitionTime":"2025-11-25T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:53 crc kubenswrapper[4802]: E1125 16:47:53.654518 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:53Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:53 crc kubenswrapper[4802]: E1125 16:47:53.654656 4802 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.656086 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.656111 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.656149 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.656174 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.656183 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:53Z","lastTransitionTime":"2025-11-25T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.758445 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.758511 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.758524 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.758541 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.758552 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:53Z","lastTransitionTime":"2025-11-25T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.862206 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.862260 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.862274 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.862295 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.862317 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:53Z","lastTransitionTime":"2025-11-25T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.965238 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.965289 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.965311 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.965338 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:53 crc kubenswrapper[4802]: I1125 16:47:53.965358 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:53Z","lastTransitionTime":"2025-11-25T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.067529 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.067567 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.067575 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.067589 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.067597 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:54Z","lastTransitionTime":"2025-11-25T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.169915 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.170203 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.170289 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.170359 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.170434 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:54Z","lastTransitionTime":"2025-11-25T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.272848 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.272880 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.272889 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.272903 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.272913 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:54Z","lastTransitionTime":"2025-11-25T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.375498 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.376437 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.376469 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.376496 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.376509 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:54Z","lastTransitionTime":"2025-11-25T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.479230 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.479307 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.479337 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.479358 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.479369 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:54Z","lastTransitionTime":"2025-11-25T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.504074 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.504196 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:54 crc kubenswrapper[4802]: E1125 16:47:54.504306 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:54 crc kubenswrapper[4802]: E1125 16:47:54.504399 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.504518 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:54 crc kubenswrapper[4802]: E1125 16:47:54.504593 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.582080 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.582167 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.582185 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.582209 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.582225 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:54Z","lastTransitionTime":"2025-11-25T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.684590 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.684629 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.684638 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.684652 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.684663 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:54Z","lastTransitionTime":"2025-11-25T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.786492 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.786557 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.786570 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.786588 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.786599 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:54Z","lastTransitionTime":"2025-11-25T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.889958 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.890006 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.890031 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.890049 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.890058 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:54Z","lastTransitionTime":"2025-11-25T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.993333 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.993389 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.993405 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.993428 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:54 crc kubenswrapper[4802]: I1125 16:47:54.993444 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:54Z","lastTransitionTime":"2025-11-25T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.096412 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.096474 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.096489 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.096506 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.096517 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:55Z","lastTransitionTime":"2025-11-25T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.198386 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.198433 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.198445 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.198464 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.198476 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:55Z","lastTransitionTime":"2025-11-25T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.300848 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.300897 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.300943 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.300960 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.300972 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:55Z","lastTransitionTime":"2025-11-25T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.403552 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.403599 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.403611 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.403628 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.403644 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:55Z","lastTransitionTime":"2025-11-25T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.504386 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:55 crc kubenswrapper[4802]: E1125 16:47:55.504774 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.505069 4802 scope.go:117] "RemoveContainer" containerID="5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977" Nov 25 16:47:55 crc kubenswrapper[4802]: E1125 16:47:55.505274 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.505820 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.505864 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.505880 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.505905 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.505922 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:55Z","lastTransitionTime":"2025-11-25T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.608159 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.608198 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.608206 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.608223 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.608233 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:55Z","lastTransitionTime":"2025-11-25T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.710741 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.710771 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.710784 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.710798 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.710809 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:55Z","lastTransitionTime":"2025-11-25T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.814604 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.814658 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.814687 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.814712 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.814729 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:55Z","lastTransitionTime":"2025-11-25T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.916761 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.916790 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.916798 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.916810 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:55 crc kubenswrapper[4802]: I1125 16:47:55.916818 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:55Z","lastTransitionTime":"2025-11-25T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.019572 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.019617 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.019626 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.019639 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.019649 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:56Z","lastTransitionTime":"2025-11-25T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.122787 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.122848 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.122860 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.122878 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.122892 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:56Z","lastTransitionTime":"2025-11-25T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.225901 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.225983 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.226009 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.226042 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.226066 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:56Z","lastTransitionTime":"2025-11-25T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.329497 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.329550 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.329563 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.329582 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.329595 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:56Z","lastTransitionTime":"2025-11-25T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.432006 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.432085 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.432110 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.432173 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.432199 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:56Z","lastTransitionTime":"2025-11-25T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.504371 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.504511 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.504371 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:56 crc kubenswrapper[4802]: E1125 16:47:56.504568 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:56 crc kubenswrapper[4802]: E1125 16:47:56.504733 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:47:56 crc kubenswrapper[4802]: E1125 16:47:56.504855 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.534649 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.534703 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.534722 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.534746 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.534765 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:56Z","lastTransitionTime":"2025-11-25T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.638272 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.638328 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.638344 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.638367 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.638386 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:56Z","lastTransitionTime":"2025-11-25T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.741166 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.741236 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.741259 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.741298 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.741350 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:56Z","lastTransitionTime":"2025-11-25T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.843652 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.843686 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.843694 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.843706 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.843715 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:56Z","lastTransitionTime":"2025-11-25T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.946682 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.946749 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.946767 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.946793 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:56 crc kubenswrapper[4802]: I1125 16:47:56.946811 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:56Z","lastTransitionTime":"2025-11-25T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.049721 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.049776 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.049791 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.049810 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.049825 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:57Z","lastTransitionTime":"2025-11-25T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.152448 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.152488 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.152500 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.152517 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.152554 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:57Z","lastTransitionTime":"2025-11-25T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.254884 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.254937 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.254956 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.254971 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.254980 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:57Z","lastTransitionTime":"2025-11-25T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.357928 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.358038 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.358070 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.358104 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.358151 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:57Z","lastTransitionTime":"2025-11-25T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.461506 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.461576 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.461594 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.461616 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.461632 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:57Z","lastTransitionTime":"2025-11-25T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.504217 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:57 crc kubenswrapper[4802]: E1125 16:47:57.504890 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.527820 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:42Z\\\",\\\"message\\\":\\\"4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1125 16:47:42.255841 6417 lb_config.go:1031] Cluster endpoints for openshift-console-operator/metrics for network=default are: map[]\\\\nF1125 16:47:42.255847 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z]\\\\nI1125 16:47:42.255857 6417 services_controller.go:443] Built service openshift-console-operator/metrics LB cluster-wide con\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:57Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.537652 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:57Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.549030 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:57Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.558780 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:57Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.563601 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.563632 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.563640 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.563652 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.563661 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:57Z","lastTransitionTime":"2025-11-25T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.570186 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:57Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.582344 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:57Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.594659 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:57Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.605883 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:57Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.618827 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:57Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.630800 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:57Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.641184 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:57Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.656962 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:57Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.665581 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.665611 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.665622 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.665637 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.665647 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:57Z","lastTransitionTime":"2025-11-25T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.671290 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:57Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.682439 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:57Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.692556 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd9aa596-76c2-468c-b732-d85bc1f16070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6ft2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:57Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.703079 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fd65faa-292f-433b-83de-043c66cf484d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://060bdec5b0ff95b1ac4dcdb60f4bbb811ddeff0c93e74e373640bf65cb1d5ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb75f902fead2f9945afbecfdfd7b7fd35c5eaa4a05059a2038fb08e07875ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4bee24ceffdebad39410dfdca5f8a83c8664f6b4c18cc2495f47f47162d116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:57Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.711736 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feaf869b-f88a-4874-8174-b689fcf6f40a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d672b4cc7df9d1855c8e882c5f4c14dec0df01a8eadab587920753fee67abf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92fbcf119fd704d01bca613b32439360422fe8ea79ba3331eace6d7f52cb5d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7p6zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:57Z is after 2025-08-24T17:21:41Z" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.767114 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.767157 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.767165 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.767179 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.767187 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:57Z","lastTransitionTime":"2025-11-25T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.869637 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.869674 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.869685 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.869700 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.869712 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:57Z","lastTransitionTime":"2025-11-25T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.971788 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.971823 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.971834 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.971849 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:57 crc kubenswrapper[4802]: I1125 16:47:57.971859 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:57Z","lastTransitionTime":"2025-11-25T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.074151 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.074214 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.074224 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.074240 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.074250 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:58Z","lastTransitionTime":"2025-11-25T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.176695 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.176756 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.176772 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.176794 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.176809 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:58Z","lastTransitionTime":"2025-11-25T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.279389 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.279455 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.279470 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.279494 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.279506 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:58Z","lastTransitionTime":"2025-11-25T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.383861 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.383899 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.383909 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.383922 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.383932 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:58Z","lastTransitionTime":"2025-11-25T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.486213 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.486257 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.486270 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.486287 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.486300 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:58Z","lastTransitionTime":"2025-11-25T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.503781 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.503797 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:47:58 crc kubenswrapper[4802]: E1125 16:47:58.503954 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:47:58 crc kubenswrapper[4802]: E1125 16:47:58.504081 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.503796 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:47:58 crc kubenswrapper[4802]: E1125 16:47:58.504213 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.589471 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.589527 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.589539 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.589556 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.589570 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:58Z","lastTransitionTime":"2025-11-25T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.692517 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.692578 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.692588 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.692605 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.692615 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:58Z","lastTransitionTime":"2025-11-25T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.795268 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.795351 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.795367 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.795408 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.795422 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:58Z","lastTransitionTime":"2025-11-25T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.897849 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.897907 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.897934 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.897957 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:58 crc kubenswrapper[4802]: I1125 16:47:58.898018 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:58Z","lastTransitionTime":"2025-11-25T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.000969 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.001008 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.001042 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.001058 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.001067 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:59Z","lastTransitionTime":"2025-11-25T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.104558 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.104600 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.104611 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.104625 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.104634 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:59Z","lastTransitionTime":"2025-11-25T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.206869 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.206935 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.206949 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.206964 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.206974 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:59Z","lastTransitionTime":"2025-11-25T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.309332 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.309375 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.309409 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.309426 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.309435 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:59Z","lastTransitionTime":"2025-11-25T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.412390 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.412457 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.412474 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.412498 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.412515 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:59Z","lastTransitionTime":"2025-11-25T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.504432 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:47:59 crc kubenswrapper[4802]: E1125 16:47:59.504644 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.514114 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.514174 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.514186 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.514203 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.514213 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:59Z","lastTransitionTime":"2025-11-25T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.616609 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.616852 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.616946 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.617021 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.617092 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:59Z","lastTransitionTime":"2025-11-25T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.719513 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.719772 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.719835 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.719937 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.720024 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:59Z","lastTransitionTime":"2025-11-25T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.822836 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.823019 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.823036 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.823055 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.823066 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:59Z","lastTransitionTime":"2025-11-25T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.925495 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.925586 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.925607 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.925630 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:47:59 crc kubenswrapper[4802]: I1125 16:47:59.925647 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:47:59Z","lastTransitionTime":"2025-11-25T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.028182 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.028458 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.028543 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.028625 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.028700 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:00Z","lastTransitionTime":"2025-11-25T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.132029 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.132171 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.132201 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.132236 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.132265 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:00Z","lastTransitionTime":"2025-11-25T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.235945 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.236515 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.236632 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.237187 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.237493 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:00Z","lastTransitionTime":"2025-11-25T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.339623 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.339659 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.339670 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.339687 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.339700 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:00Z","lastTransitionTime":"2025-11-25T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.442648 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.442681 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.442692 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.442705 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.442716 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:00Z","lastTransitionTime":"2025-11-25T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.504414 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.504473 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.504517 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:00 crc kubenswrapper[4802]: E1125 16:48:00.504778 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:00 crc kubenswrapper[4802]: E1125 16:48:00.504978 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:00 crc kubenswrapper[4802]: E1125 16:48:00.505295 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.545171 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.545221 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.545231 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.545249 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.545261 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:00Z","lastTransitionTime":"2025-11-25T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.647807 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.647937 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.647965 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.647989 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.648008 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:00Z","lastTransitionTime":"2025-11-25T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.751002 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.751058 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.751075 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.751097 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.751110 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:00Z","lastTransitionTime":"2025-11-25T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.853488 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.853548 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.853558 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.853592 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.853603 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:00Z","lastTransitionTime":"2025-11-25T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.955543 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.955581 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.955592 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.955608 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:00 crc kubenswrapper[4802]: I1125 16:48:00.955619 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:00Z","lastTransitionTime":"2025-11-25T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.058547 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.058829 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.058901 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.058982 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.059074 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:01Z","lastTransitionTime":"2025-11-25T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.161458 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.161496 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.161507 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.161523 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.161535 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:01Z","lastTransitionTime":"2025-11-25T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.263493 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.263525 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.263532 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.263545 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.263555 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:01Z","lastTransitionTime":"2025-11-25T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.367217 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.367280 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.367291 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.367312 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.367325 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:01Z","lastTransitionTime":"2025-11-25T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.471031 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.471102 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.471159 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.471193 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.471217 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:01Z","lastTransitionTime":"2025-11-25T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.503759 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:01 crc kubenswrapper[4802]: E1125 16:48:01.504313 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.574943 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.575028 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.575054 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.575154 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.575189 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:01Z","lastTransitionTime":"2025-11-25T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.677773 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.677819 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.677830 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.677851 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.677865 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:01Z","lastTransitionTime":"2025-11-25T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.779983 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.780029 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.780040 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.780060 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.780074 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:01Z","lastTransitionTime":"2025-11-25T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.887698 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.887738 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.887750 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.887768 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.887778 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:01Z","lastTransitionTime":"2025-11-25T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.990418 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.990477 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.990493 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.990522 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:01 crc kubenswrapper[4802]: I1125 16:48:01.990544 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:01Z","lastTransitionTime":"2025-11-25T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.092518 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.092557 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.092567 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.092583 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.092595 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:02Z","lastTransitionTime":"2025-11-25T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.194843 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.195114 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.195238 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.195337 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.195379 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:02Z","lastTransitionTime":"2025-11-25T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.298012 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.298069 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.298081 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.298093 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.298101 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:02Z","lastTransitionTime":"2025-11-25T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.400390 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.400469 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.400539 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.400571 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.400593 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:02Z","lastTransitionTime":"2025-11-25T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.502948 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.503000 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.503012 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.503031 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.503045 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:02Z","lastTransitionTime":"2025-11-25T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.503789 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.503825 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.503832 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:02 crc kubenswrapper[4802]: E1125 16:48:02.503888 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:02 crc kubenswrapper[4802]: E1125 16:48:02.504014 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:02 crc kubenswrapper[4802]: E1125 16:48:02.504096 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.605342 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.605380 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.605392 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.605408 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.605420 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:02Z","lastTransitionTime":"2025-11-25T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.708000 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.708027 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.708035 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.708047 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.708056 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:02Z","lastTransitionTime":"2025-11-25T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.811007 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.811079 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.811102 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.811162 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.811184 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:02Z","lastTransitionTime":"2025-11-25T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.913761 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.913811 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.913820 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.913835 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:02 crc kubenswrapper[4802]: I1125 16:48:02.913844 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:02Z","lastTransitionTime":"2025-11-25T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.017193 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.017244 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.017254 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.017282 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.017323 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:03Z","lastTransitionTime":"2025-11-25T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.120775 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.120836 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.120850 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.120865 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.120874 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:03Z","lastTransitionTime":"2025-11-25T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.223282 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.223364 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.223442 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.223474 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.223496 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:03Z","lastTransitionTime":"2025-11-25T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.325460 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.325515 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.325534 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.325559 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.325576 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:03Z","lastTransitionTime":"2025-11-25T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.427555 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.427611 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.427634 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.427691 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.427716 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:03Z","lastTransitionTime":"2025-11-25T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.504562 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:03 crc kubenswrapper[4802]: E1125 16:48:03.504678 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.529715 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.529742 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.529750 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.529763 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.529772 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:03Z","lastTransitionTime":"2025-11-25T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.631788 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.631832 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.631843 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.631861 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.631874 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:03Z","lastTransitionTime":"2025-11-25T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.734111 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.734168 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.734176 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.734190 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.734199 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:03Z","lastTransitionTime":"2025-11-25T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.836214 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.836245 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.836256 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.836271 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.836281 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:03Z","lastTransitionTime":"2025-11-25T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.938774 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.938801 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.938810 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.938823 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.938833 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:03Z","lastTransitionTime":"2025-11-25T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.966612 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.966651 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.966660 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.966677 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.966686 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:03Z","lastTransitionTime":"2025-11-25T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:03 crc kubenswrapper[4802]: E1125 16:48:03.985196 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:03Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.989240 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.989302 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.989313 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.989334 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:03 crc kubenswrapper[4802]: I1125 16:48:03.989346 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:03Z","lastTransitionTime":"2025-11-25T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:04 crc kubenswrapper[4802]: E1125 16:48:04.004957 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:04Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.011648 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.011838 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.011951 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.012072 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.012201 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:04Z","lastTransitionTime":"2025-11-25T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:04 crc kubenswrapper[4802]: E1125 16:48:04.027335 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:04Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.032399 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.032457 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.032467 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.032488 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.032500 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:04Z","lastTransitionTime":"2025-11-25T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:04 crc kubenswrapper[4802]: E1125 16:48:04.044758 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:04Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.049431 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.049490 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.049508 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.049530 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.049545 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:04Z","lastTransitionTime":"2025-11-25T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:04 crc kubenswrapper[4802]: E1125 16:48:04.063602 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:04Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:04 crc kubenswrapper[4802]: E1125 16:48:04.063769 4802 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.066326 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.066368 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.066378 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.066397 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.066413 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:04Z","lastTransitionTime":"2025-11-25T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.169508 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.169548 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.169568 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.169585 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.169596 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:04Z","lastTransitionTime":"2025-11-25T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.271913 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.271986 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.272000 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.272019 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.272058 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:04Z","lastTransitionTime":"2025-11-25T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.375251 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.375300 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.375312 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.375332 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.375343 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:04Z","lastTransitionTime":"2025-11-25T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.477591 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.477863 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.477878 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.477909 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.477924 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:04Z","lastTransitionTime":"2025-11-25T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.504122 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.504313 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:04 crc kubenswrapper[4802]: E1125 16:48:04.504403 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.504441 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:04 crc kubenswrapper[4802]: E1125 16:48:04.504604 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:04 crc kubenswrapper[4802]: E1125 16:48:04.504728 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.567059 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs\") pod \"network-metrics-daemon-6ft2z\" (UID: \"cd9aa596-76c2-468c-b732-d85bc1f16070\") " pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:04 crc kubenswrapper[4802]: E1125 16:48:04.567225 4802 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 16:48:04 crc kubenswrapper[4802]: E1125 16:48:04.567286 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs podName:cd9aa596-76c2-468c-b732-d85bc1f16070 nodeName:}" failed. No retries permitted until 2025-11-25 16:48:36.567268219 +0000 UTC m=+99.711615405 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs") pod "network-metrics-daemon-6ft2z" (UID: "cd9aa596-76c2-468c-b732-d85bc1f16070") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.579875 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.579901 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.579908 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.579921 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.579932 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:04Z","lastTransitionTime":"2025-11-25T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.683094 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.683151 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.683162 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.683179 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.683191 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:04Z","lastTransitionTime":"2025-11-25T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.785928 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.785971 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.785984 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.786009 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.786021 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:04Z","lastTransitionTime":"2025-11-25T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.888302 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.888365 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.888378 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.888395 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.888406 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:04Z","lastTransitionTime":"2025-11-25T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.990670 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.990713 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.990742 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.990757 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:04 crc kubenswrapper[4802]: I1125 16:48:04.990766 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:04Z","lastTransitionTime":"2025-11-25T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.092811 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.092883 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.092906 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.092935 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.092962 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:05Z","lastTransitionTime":"2025-11-25T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.196623 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.196694 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.196705 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.196718 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.196728 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:05Z","lastTransitionTime":"2025-11-25T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.298537 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.298580 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.298589 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.298609 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.298619 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:05Z","lastTransitionTime":"2025-11-25T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.401359 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.401398 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.401407 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.401422 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.401431 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:05Z","lastTransitionTime":"2025-11-25T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.503101 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.503185 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.503200 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.503217 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.503228 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:05Z","lastTransitionTime":"2025-11-25T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.503491 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:05 crc kubenswrapper[4802]: E1125 16:48:05.503578 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.605021 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.605058 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.605089 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.605105 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.605115 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:05Z","lastTransitionTime":"2025-11-25T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.707924 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.707961 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.707974 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.707989 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.707999 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:05Z","lastTransitionTime":"2025-11-25T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.809935 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.809959 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.809987 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.810002 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.810012 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:05Z","lastTransitionTime":"2025-11-25T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.912322 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.912379 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.912392 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.912412 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:05 crc kubenswrapper[4802]: I1125 16:48:05.912425 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:05Z","lastTransitionTime":"2025-11-25T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.015101 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.015129 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.015138 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.015164 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.015173 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:06Z","lastTransitionTime":"2025-11-25T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.117447 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.117685 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.117776 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.117839 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.117924 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:06Z","lastTransitionTime":"2025-11-25T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.220496 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.220529 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.220538 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.220550 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.220561 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:06Z","lastTransitionTime":"2025-11-25T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.322753 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.322788 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.322801 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.322820 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.322832 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:06Z","lastTransitionTime":"2025-11-25T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.425425 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.425468 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.425478 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.425499 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.425514 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:06Z","lastTransitionTime":"2025-11-25T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.503614 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.503636 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.503697 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:06 crc kubenswrapper[4802]: E1125 16:48:06.503773 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:06 crc kubenswrapper[4802]: E1125 16:48:06.503896 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:06 crc kubenswrapper[4802]: E1125 16:48:06.503998 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.528964 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.529003 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.529011 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.529026 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.529035 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:06Z","lastTransitionTime":"2025-11-25T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.631606 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.631641 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.631650 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.631664 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.631674 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:06Z","lastTransitionTime":"2025-11-25T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.734350 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.734392 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.734400 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.734428 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.734438 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:06Z","lastTransitionTime":"2025-11-25T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.818641 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hmbvn_97e822d6-58fe-41f4-b08b-3c9b42273307/kube-multus/0.log" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.818688 4802 generic.go:334] "Generic (PLEG): container finished" podID="97e822d6-58fe-41f4-b08b-3c9b42273307" containerID="ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b" exitCode=1 Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.818715 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hmbvn" event={"ID":"97e822d6-58fe-41f4-b08b-3c9b42273307","Type":"ContainerDied","Data":"ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b"} Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.819077 4802 scope.go:117] "RemoveContainer" containerID="ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.837080 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.837305 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.837313 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.837327 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.837337 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:06Z","lastTransitionTime":"2025-11-25T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.846107 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:42Z\\\",\\\"message\\\":\\\"4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1125 16:47:42.255841 6417 lb_config.go:1031] Cluster endpoints for openshift-console-operator/metrics for network=default are: map[]\\\\nF1125 16:47:42.255847 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z]\\\\nI1125 16:47:42.255857 6417 services_controller.go:443] Built service openshift-console-operator/metrics LB cluster-wide con\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:06Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.857266 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:06Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.868349 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:06Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.881114 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:06Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.895577 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:06Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.904235 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:06Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.914236 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:06Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.929238 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:06Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.939687 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.939730 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.939739 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.939754 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.939763 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:06Z","lastTransitionTime":"2025-11-25T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.948556 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:06Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.961805 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:06Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.975658 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:06Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:06 crc kubenswrapper[4802]: I1125 16:48:06.991376 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:06Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.002550 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.013837 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:48:06Z\\\",\\\"message\\\":\\\"2025-11-25T16:47:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_07bab96e-be78-4d51-aab3-a524db7619cd\\\\n2025-11-25T16:47:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_07bab96e-be78-4d51-aab3-a524db7619cd to /host/opt/cni/bin/\\\\n2025-11-25T16:47:21Z [verbose] multus-daemon started\\\\n2025-11-25T16:47:21Z [verbose] Readiness Indicator file check\\\\n2025-11-25T16:48:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.022635 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd9aa596-76c2-468c-b732-d85bc1f16070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6ft2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.032280 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fd65faa-292f-433b-83de-043c66cf484d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://060bdec5b0ff95b1ac4dcdb60f4bbb811ddeff0c93e74e373640bf65cb1d5ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb75f902fead2f9945afbecfdfd7b7fd35c5eaa4a05059a2038fb08e07875ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4bee24ceffdebad39410dfdca5f8a83c8664f6b4c18cc2495f47f47162d116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.042047 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.042091 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.042099 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.042113 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.042125 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:07Z","lastTransitionTime":"2025-11-25T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.043316 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feaf869b-f88a-4874-8174-b689fcf6f40a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d672b4cc7df9d1855c8e882c5f4c14dec0df01a8eadab587920753fee67abf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92fbcf119fd704d01bca613b32439360422fe8ea79ba3331eace6d7f52cb5d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7p6zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.144966 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.145010 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.145022 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.145044 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.145057 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:07Z","lastTransitionTime":"2025-11-25T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.247480 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.247766 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.247886 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.247981 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.248065 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:07Z","lastTransitionTime":"2025-11-25T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.352240 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.352315 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.352325 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.352345 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.352356 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:07Z","lastTransitionTime":"2025-11-25T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.454951 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.455033 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.455053 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.455095 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.455111 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:07Z","lastTransitionTime":"2025-11-25T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.504401 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:07 crc kubenswrapper[4802]: E1125 16:48:07.504635 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.537895 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:42Z\\\",\\\"message\\\":\\\"4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1125 16:47:42.255841 6417 lb_config.go:1031] Cluster endpoints for openshift-console-operator/metrics for network=default are: map[]\\\\nF1125 16:47:42.255847 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z]\\\\nI1125 16:47:42.255857 6417 services_controller.go:443] Built service openshift-console-operator/metrics LB cluster-wide con\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.554486 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.560991 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.561032 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.561042 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.561057 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.561074 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:07Z","lastTransitionTime":"2025-11-25T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.571850 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.586414 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.603788 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.618364 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.633685 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.650098 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.664954 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.665032 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.665050 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.665076 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.665092 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:07Z","lastTransitionTime":"2025-11-25T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.668840 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.686917 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.702094 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.717968 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.731286 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.745300 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:48:06Z\\\",\\\"message\\\":\\\"2025-11-25T16:47:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_07bab96e-be78-4d51-aab3-a524db7619cd\\\\n2025-11-25T16:47:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_07bab96e-be78-4d51-aab3-a524db7619cd to /host/opt/cni/bin/\\\\n2025-11-25T16:47:21Z [verbose] multus-daemon started\\\\n2025-11-25T16:47:21Z [verbose] Readiness Indicator file check\\\\n2025-11-25T16:48:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.759497 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd9aa596-76c2-468c-b732-d85bc1f16070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6ft2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.768173 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.768213 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.768225 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.768241 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.768251 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:07Z","lastTransitionTime":"2025-11-25T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.772824 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fd65faa-292f-433b-83de-043c66cf484d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://060bdec5b0ff95b1ac4dcdb60f4bbb811ddeff0c93e74e373640bf65cb1d5ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb75f902fead2f9945afbecfdfd7b7fd35c5eaa4a05059a2038fb08e07875ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4bee24ceffdebad39410dfdca5f8a83c8664f6b4c18cc2495f47f47162d116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.783715 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feaf869b-f88a-4874-8174-b689fcf6f40a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d672b4cc7df9d1855c8e882c5f4c14dec0df01a8eadab587920753fee67abf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92fbcf119fd704d01bca613b32439360422fe8ea79ba3331eace6d7f52cb5d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7p6zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.823217 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hmbvn_97e822d6-58fe-41f4-b08b-3c9b42273307/kube-multus/0.log" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.823277 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hmbvn" event={"ID":"97e822d6-58fe-41f4-b08b-3c9b42273307","Type":"ContainerStarted","Data":"e0e948183d9361bf5f6831d25601431a1de0662987928a7277917451064c0295"} Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.837095 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.856419 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.868160 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.870849 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.870925 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.870936 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.870950 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.870959 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:07Z","lastTransitionTime":"2025-11-25T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.880420 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.893359 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0e948183d9361bf5f6831d25601431a1de0662987928a7277917451064c0295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:48:06Z\\\",\\\"message\\\":\\\"2025-11-25T16:47:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_07bab96e-be78-4d51-aab3-a524db7619cd\\\\n2025-11-25T16:47:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_07bab96e-be78-4d51-aab3-a524db7619cd to /host/opt/cni/bin/\\\\n2025-11-25T16:47:21Z [verbose] multus-daemon started\\\\n2025-11-25T16:47:21Z [verbose] Readiness Indicator file check\\\\n2025-11-25T16:48:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.904409 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.923528 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.941974 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.954874 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd9aa596-76c2-468c-b732-d85bc1f16070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6ft2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.968810 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feaf869b-f88a-4874-8174-b689fcf6f40a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d672b4cc7df9d1855c8e882c5f4c14dec0df01a8eadab587920753fee67abf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92fbcf119fd704d01bca613b32439360422fe8ea79ba3331eace6d7f52cb5d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7p6zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.972784 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.972810 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.972818 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.972833 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.972844 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:07Z","lastTransitionTime":"2025-11-25T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.979487 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fd65faa-292f-433b-83de-043c66cf484d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://060bdec5b0ff95b1ac4dcdb60f4bbb811ddeff0c93e74e373640bf65cb1d5ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb75f902fead2f9945afbecfdfd7b7fd35c5eaa4a05059a2038fb08e07875ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4bee24ceffdebad39410dfdca5f8a83c8664f6b4c18cc2495f47f47162d116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:07 crc kubenswrapper[4802]: I1125 16:48:07.996328 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:42Z\\\",\\\"message\\\":\\\"4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1125 16:47:42.255841 6417 lb_config.go:1031] Cluster endpoints for openshift-console-operator/metrics for network=default are: map[]\\\\nF1125 16:47:42.255847 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z]\\\\nI1125 16:47:42.255857 6417 services_controller.go:443] Built service openshift-console-operator/metrics LB cluster-wide con\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:07Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.006421 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:08Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.018711 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:08Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.030793 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:08Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.041322 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:08Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.053869 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:08Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.076380 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.076415 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.076423 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.076442 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.076453 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:08Z","lastTransitionTime":"2025-11-25T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.179324 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.179368 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.179380 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.179400 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.179411 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:08Z","lastTransitionTime":"2025-11-25T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.281010 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.281085 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.281100 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.281115 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.281145 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:08Z","lastTransitionTime":"2025-11-25T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.384412 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.384467 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.384521 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.384544 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.384597 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:08Z","lastTransitionTime":"2025-11-25T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.487469 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.487509 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.487518 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.487532 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.487541 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:08Z","lastTransitionTime":"2025-11-25T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.504379 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.504404 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.504379 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:08 crc kubenswrapper[4802]: E1125 16:48:08.504496 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:08 crc kubenswrapper[4802]: E1125 16:48:08.504579 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:08 crc kubenswrapper[4802]: E1125 16:48:08.504702 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.590339 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.590378 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.590388 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.590402 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.590412 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:08Z","lastTransitionTime":"2025-11-25T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.692597 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.692644 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.692656 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.692672 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.692681 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:08Z","lastTransitionTime":"2025-11-25T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.795109 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.795156 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.795164 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.795180 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.795197 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:08Z","lastTransitionTime":"2025-11-25T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.898107 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.898174 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.898182 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.898195 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:08 crc kubenswrapper[4802]: I1125 16:48:08.898203 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:08Z","lastTransitionTime":"2025-11-25T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.000451 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.000490 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.000499 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.000519 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.000532 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:09Z","lastTransitionTime":"2025-11-25T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.102673 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.102729 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.102743 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.102758 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.102767 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:09Z","lastTransitionTime":"2025-11-25T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.205308 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.205352 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.205366 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.205381 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.205390 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:09Z","lastTransitionTime":"2025-11-25T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.307805 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.307852 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.307860 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.307876 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.307889 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:09Z","lastTransitionTime":"2025-11-25T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.410099 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.410164 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.410175 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.410191 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.410202 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:09Z","lastTransitionTime":"2025-11-25T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.503686 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:09 crc kubenswrapper[4802]: E1125 16:48:09.503824 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.512516 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.512550 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.512557 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.512572 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.512582 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:09Z","lastTransitionTime":"2025-11-25T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.615188 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.615236 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.615255 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.615275 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.615291 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:09Z","lastTransitionTime":"2025-11-25T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.717850 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.717885 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.717896 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.717910 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.717920 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:09Z","lastTransitionTime":"2025-11-25T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.819772 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.819823 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.819835 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.819854 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.819866 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:09Z","lastTransitionTime":"2025-11-25T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.921704 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.921754 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.921766 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.921783 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:09 crc kubenswrapper[4802]: I1125 16:48:09.921794 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:09Z","lastTransitionTime":"2025-11-25T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.023477 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.023522 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.023534 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.023550 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.023565 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:10Z","lastTransitionTime":"2025-11-25T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.125811 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.125843 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.125852 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.125866 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.125875 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:10Z","lastTransitionTime":"2025-11-25T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.228161 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.228199 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.228215 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.228248 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.228266 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:10Z","lastTransitionTime":"2025-11-25T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.330699 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.330734 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.330743 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.330758 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.330768 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:10Z","lastTransitionTime":"2025-11-25T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.432567 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.432612 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.432625 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.432652 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.432667 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:10Z","lastTransitionTime":"2025-11-25T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.504428 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.504458 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.504537 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:10 crc kubenswrapper[4802]: E1125 16:48:10.505016 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:10 crc kubenswrapper[4802]: E1125 16:48:10.505205 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:10 crc kubenswrapper[4802]: E1125 16:48:10.505290 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.505390 4802 scope.go:117] "RemoveContainer" containerID="5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.534775 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.534806 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.534816 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.534831 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.534842 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:10Z","lastTransitionTime":"2025-11-25T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.636829 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.636875 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.636888 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.636905 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.636916 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:10Z","lastTransitionTime":"2025-11-25T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.739059 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.739096 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.739107 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.739143 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.739156 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:10Z","lastTransitionTime":"2025-11-25T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.835669 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovnkube-controller/2.log" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.838036 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerStarted","Data":"88787834da7629acd6521f977ee73dec4231d0b4de6a3bb8f4dad83d203b4221"} Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.838464 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.841325 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.841373 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.841382 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.841408 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.841419 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:10Z","lastTransitionTime":"2025-11-25T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.857933 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:10Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.880273 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:10Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.900535 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:10Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.913779 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:10Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.928311 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:10Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.939755 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:10Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.943190 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.943223 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.943233 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.943247 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.943256 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:10Z","lastTransitionTime":"2025-11-25T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.951099 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:10Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.961670 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0e948183d9361bf5f6831d25601431a1de0662987928a7277917451064c0295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:48:06Z\\\",\\\"message\\\":\\\"2025-11-25T16:47:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_07bab96e-be78-4d51-aab3-a524db7619cd\\\\n2025-11-25T16:47:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_07bab96e-be78-4d51-aab3-a524db7619cd to /host/opt/cni/bin/\\\\n2025-11-25T16:47:21Z [verbose] multus-daemon started\\\\n2025-11-25T16:47:21Z [verbose] Readiness Indicator file check\\\\n2025-11-25T16:48:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:10Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.971774 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd9aa596-76c2-468c-b732-d85bc1f16070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6ft2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:10Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.983071 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fd65faa-292f-433b-83de-043c66cf484d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://060bdec5b0ff95b1ac4dcdb60f4bbb811ddeff0c93e74e373640bf65cb1d5ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb75f902fead2f9945afbecfdfd7b7fd35c5eaa4a05059a2038fb08e07875ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4bee24ceffdebad39410dfdca5f8a83c8664f6b4c18cc2495f47f47162d116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:10Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:10 crc kubenswrapper[4802]: I1125 16:48:10.999336 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feaf869b-f88a-4874-8174-b689fcf6f40a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d672b4cc7df9d1855c8e882c5f4c14dec0df01a8eadab587920753fee67abf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92fbcf119fd704d01bca613b32439360422fe8ea79ba3331eace6d7f52cb5d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7p6zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:10Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.019921 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88787834da7629acd6521f977ee73dec4231d0b4de6a3bb8f4dad83d203b4221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:42Z\\\",\\\"message\\\":\\\"4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1125 16:47:42.255841 6417 lb_config.go:1031] Cluster endpoints for openshift-console-operator/metrics for network=default are: map[]\\\\nF1125 16:47:42.255847 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z]\\\\nI1125 16:47:42.255857 6417 services_controller.go:443] Built service openshift-console-operator/metrics LB cluster-wide con\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:11Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.031343 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:11Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.043339 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:11Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.045176 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.045207 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.045216 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.045230 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.045239 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:11Z","lastTransitionTime":"2025-11-25T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.057224 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:11Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.071534 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:11Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.085622 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:11Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.147185 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.147227 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.147238 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.147255 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.147266 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:11Z","lastTransitionTime":"2025-11-25T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.250066 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.250113 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.250143 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.250159 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.250168 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:11Z","lastTransitionTime":"2025-11-25T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.352235 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.352287 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.352298 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.352313 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.352322 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:11Z","lastTransitionTime":"2025-11-25T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.454906 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.454947 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.454955 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.454969 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.454979 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:11Z","lastTransitionTime":"2025-11-25T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.504346 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:11 crc kubenswrapper[4802]: E1125 16:48:11.504524 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.557631 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.557694 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.557707 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.557734 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.557749 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:11Z","lastTransitionTime":"2025-11-25T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.661035 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.661086 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.661104 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.661170 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.661188 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:11Z","lastTransitionTime":"2025-11-25T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.764262 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.764292 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.764300 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.764314 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.764325 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:11Z","lastTransitionTime":"2025-11-25T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.842956 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovnkube-controller/3.log" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.843487 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovnkube-controller/2.log" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.846009 4802 generic.go:334] "Generic (PLEG): container finished" podID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerID="88787834da7629acd6521f977ee73dec4231d0b4de6a3bb8f4dad83d203b4221" exitCode=1 Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.846050 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerDied","Data":"88787834da7629acd6521f977ee73dec4231d0b4de6a3bb8f4dad83d203b4221"} Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.846081 4802 scope.go:117] "RemoveContainer" containerID="5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.846780 4802 scope.go:117] "RemoveContainer" containerID="88787834da7629acd6521f977ee73dec4231d0b4de6a3bb8f4dad83d203b4221" Nov 25 16:48:11 crc kubenswrapper[4802]: E1125 16:48:11.846936 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.873907 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feaf869b-f88a-4874-8174-b689fcf6f40a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d672b4cc7df9d1855c8e882c5f4c14dec0df01a8eadab587920753fee67abf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92fbcf119fd704d01bca613b32439360422fe8ea79ba3331eace6d7f52cb5d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7p6zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:11Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.895657 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.895696 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.895705 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.895718 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.895727 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:11Z","lastTransitionTime":"2025-11-25T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.901414 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fd65faa-292f-433b-83de-043c66cf484d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://060bdec5b0ff95b1ac4dcdb60f4bbb811ddeff0c93e74e373640bf65cb1d5ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb75f902fead2f9945afbecfdfd7b7fd35c5eaa4a05059a2038fb08e07875ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4bee24ceffdebad39410dfdca5f8a83c8664f6b4c18cc2495f47f47162d116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:11Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.927693 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88787834da7629acd6521f977ee73dec4231d0b4de6a3bb8f4dad83d203b4221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc6b708d6f11ed4a9f1582d37aadb2e1b8f8f6879ba6fac2d95f383c3e20977\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:47:42Z\\\",\\\"message\\\":\\\"4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1125 16:47:42.255841 6417 lb_config.go:1031] Cluster endpoints for openshift-console-operator/metrics for network=default are: map[]\\\\nF1125 16:47:42.255847 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:47:42Z is after 2025-08-24T17:21:41Z]\\\\nI1125 16:47:42.255857 6417 services_controller.go:443] Built service openshift-console-operator/metrics LB cluster-wide con\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88787834da7629acd6521f977ee73dec4231d0b4de6a3bb8f4dad83d203b4221\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:48:11Z\\\",\\\"message\\\":\\\"column _uuid == {d8772e82-b0a4-4596-87d3-3d517c13344b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 16:48:11.274563 6791 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1125 16:48:11.274603 6791 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed call\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:48:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:11Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.939158 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:11Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.951614 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:11Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.965550 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:11Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.976387 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:11Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.992615 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:11Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.997939 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.997971 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.997980 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.997993 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:11 crc kubenswrapper[4802]: I1125 16:48:11.998002 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:11Z","lastTransitionTime":"2025-11-25T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.010404 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:12Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.029733 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:12Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.049486 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:12Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.061812 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:12Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.073941 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0e948183d9361bf5f6831d25601431a1de0662987928a7277917451064c0295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:48:06Z\\\",\\\"message\\\":\\\"2025-11-25T16:47:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_07bab96e-be78-4d51-aab3-a524db7619cd\\\\n2025-11-25T16:47:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_07bab96e-be78-4d51-aab3-a524db7619cd to /host/opt/cni/bin/\\\\n2025-11-25T16:47:21Z [verbose] multus-daemon started\\\\n2025-11-25T16:47:21Z [verbose] Readiness Indicator file check\\\\n2025-11-25T16:48:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:12Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.087686 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:12Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.099590 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.099626 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.099636 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.099652 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.099663 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:12Z","lastTransitionTime":"2025-11-25T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.103737 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:12Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.120412 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:12Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.131279 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd9aa596-76c2-468c-b732-d85bc1f16070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6ft2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:12Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.201680 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.201732 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.201742 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.201758 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.201768 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:12Z","lastTransitionTime":"2025-11-25T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.304680 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.304743 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.304768 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.304798 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.304823 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:12Z","lastTransitionTime":"2025-11-25T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.407515 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.407562 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.407571 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.407586 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.407595 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:12Z","lastTransitionTime":"2025-11-25T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.503602 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.503644 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.503709 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:12 crc kubenswrapper[4802]: E1125 16:48:12.503842 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:12 crc kubenswrapper[4802]: E1125 16:48:12.503952 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:12 crc kubenswrapper[4802]: E1125 16:48:12.504146 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.510583 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.510619 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.510629 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.510643 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.510651 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:12Z","lastTransitionTime":"2025-11-25T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.514017 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.613453 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.613506 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.613523 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.613546 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.613563 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:12Z","lastTransitionTime":"2025-11-25T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.716818 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.716849 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.716860 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.716877 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.716888 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:12Z","lastTransitionTime":"2025-11-25T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.820340 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.820376 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.820386 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.820404 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.820415 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:12Z","lastTransitionTime":"2025-11-25T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.851806 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovnkube-controller/3.log" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.859062 4802 scope.go:117] "RemoveContainer" containerID="88787834da7629acd6521f977ee73dec4231d0b4de6a3bb8f4dad83d203b4221" Nov 25 16:48:12 crc kubenswrapper[4802]: E1125 16:48:12.859366 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.881707 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:12Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.900045 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:12Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.920568 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:12Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.923020 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.923086 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.923110 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.923189 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.923228 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:12Z","lastTransitionTime":"2025-11-25T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.934456 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:12Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.948639 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:12Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.970402 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0e948183d9361bf5f6831d25601431a1de0662987928a7277917451064c0295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:48:06Z\\\",\\\"message\\\":\\\"2025-11-25T16:47:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_07bab96e-be78-4d51-aab3-a524db7619cd\\\\n2025-11-25T16:47:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_07bab96e-be78-4d51-aab3-a524db7619cd to /host/opt/cni/bin/\\\\n2025-11-25T16:47:21Z [verbose] multus-daemon started\\\\n2025-11-25T16:47:21Z [verbose] Readiness Indicator file check\\\\n2025-11-25T16:48:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:12Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:12 crc kubenswrapper[4802]: I1125 16:48:12.985832 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:12Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.000884 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:12Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.016075 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:13Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.026367 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.026419 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.026430 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.026447 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.026459 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:13Z","lastTransitionTime":"2025-11-25T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.029336 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:13Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.041772 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:13Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.053099 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:13Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.063028 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd9aa596-76c2-468c-b732-d85bc1f16070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6ft2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:13Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.075230 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fd65faa-292f-433b-83de-043c66cf484d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://060bdec5b0ff95b1ac4dcdb60f4bbb811ddeff0c93e74e373640bf65cb1d5ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb75f902fead2f9945afbecfdfd7b7fd35c5eaa4a05059a2038fb08e07875ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4bee24ceffdebad39410dfdca5f8a83c8664f6b4c18cc2495f47f47162d116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:13Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.085426 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feaf869b-f88a-4874-8174-b689fcf6f40a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d672b4cc7df9d1855c8e882c5f4c14dec0df01a8eadab587920753fee67abf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92fbcf119fd704d01bca613b32439360422fe8ea79ba3331eace6d7f52cb5d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7p6zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:13Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.094162 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b4845-01e3-4ae3-9027-b6ffe66c4bc6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db167a028aae7e1be86605fcfb5a716b35c6e34d600b9675fb2c26b88bc52eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74e0f0d4a5a5b635060b11f8b506df6000fe73f589125d64bdc0450dec47ccbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74e0f0d4a5a5b635060b11f8b506df6000fe73f589125d64bdc0450dec47ccbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:13Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.110501 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88787834da7629acd6521f977ee73dec4231d0b4de6a3bb8f4dad83d203b4221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88787834da7629acd6521f977ee73dec4231d0b4de6a3bb8f4dad83d203b4221\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:48:11Z\\\",\\\"message\\\":\\\"column _uuid == {d8772e82-b0a4-4596-87d3-3d517c13344b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 16:48:11.274563 6791 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1125 16:48:11.274603 6791 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed call\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:48:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:13Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.121468 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:13Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.128988 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.129018 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.129026 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.129039 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.129048 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:13Z","lastTransitionTime":"2025-11-25T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.230902 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.230938 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.230946 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.230977 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.230986 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:13Z","lastTransitionTime":"2025-11-25T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.333383 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.333416 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.333427 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.333441 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.333452 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:13Z","lastTransitionTime":"2025-11-25T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.435884 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.435949 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.435965 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.435988 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.436005 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:13Z","lastTransitionTime":"2025-11-25T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.503842 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:13 crc kubenswrapper[4802]: E1125 16:48:13.504036 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.538170 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.538225 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.538238 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.538255 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.538267 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:13Z","lastTransitionTime":"2025-11-25T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.640745 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.640832 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.640851 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.640866 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.640892 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:13Z","lastTransitionTime":"2025-11-25T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.743854 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.743930 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.743973 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.744027 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.744056 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:13Z","lastTransitionTime":"2025-11-25T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.846805 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.846840 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.846857 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.846872 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.846881 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:13Z","lastTransitionTime":"2025-11-25T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.948587 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.948629 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.948638 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.948651 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:13 crc kubenswrapper[4802]: I1125 16:48:13.948662 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:13Z","lastTransitionTime":"2025-11-25T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.050638 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.050673 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.050685 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.050700 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.050713 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:14Z","lastTransitionTime":"2025-11-25T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.152763 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.152834 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.152847 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.152865 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.152878 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:14Z","lastTransitionTime":"2025-11-25T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.209684 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.209718 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.209726 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.209742 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.209752 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:14Z","lastTransitionTime":"2025-11-25T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:14 crc kubenswrapper[4802]: E1125 16:48:14.225729 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:14Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.230169 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.230204 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.230212 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.230224 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.230234 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:14Z","lastTransitionTime":"2025-11-25T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:14 crc kubenswrapper[4802]: E1125 16:48:14.241445 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:14Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.245282 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.245351 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.245370 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.245393 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.245410 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:14Z","lastTransitionTime":"2025-11-25T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:14 crc kubenswrapper[4802]: E1125 16:48:14.262052 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:14Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.266395 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.266453 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.266467 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.266487 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.266504 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:14Z","lastTransitionTime":"2025-11-25T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:14 crc kubenswrapper[4802]: E1125 16:48:14.281227 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:14Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.285154 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.285204 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.285216 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.285236 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.285247 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:14Z","lastTransitionTime":"2025-11-25T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:14 crc kubenswrapper[4802]: E1125 16:48:14.300270 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:14Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:14 crc kubenswrapper[4802]: E1125 16:48:14.300425 4802 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.302439 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.302485 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.302497 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.302518 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.302536 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:14Z","lastTransitionTime":"2025-11-25T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.404841 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.404889 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.404899 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.404916 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.404929 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:14Z","lastTransitionTime":"2025-11-25T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.503762 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.503849 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:14 crc kubenswrapper[4802]: E1125 16:48:14.503885 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.503774 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:14 crc kubenswrapper[4802]: E1125 16:48:14.503997 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:14 crc kubenswrapper[4802]: E1125 16:48:14.504094 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.507096 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.507147 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.507157 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.507170 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.507180 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:14Z","lastTransitionTime":"2025-11-25T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.610219 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.610272 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.610283 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.610301 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.610313 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:14Z","lastTransitionTime":"2025-11-25T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.712723 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.712764 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.712774 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.712790 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.712801 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:14Z","lastTransitionTime":"2025-11-25T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.816034 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.816072 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.816081 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.816093 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.816102 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:14Z","lastTransitionTime":"2025-11-25T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.918737 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.918773 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.918785 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.918803 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:14 crc kubenswrapper[4802]: I1125 16:48:14.918814 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:14Z","lastTransitionTime":"2025-11-25T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.021017 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.021044 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.021053 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.021067 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.021076 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:15Z","lastTransitionTime":"2025-11-25T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.123020 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.123054 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.123065 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.123079 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.123088 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:15Z","lastTransitionTime":"2025-11-25T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.225415 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.225442 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.225451 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.225463 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.225472 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:15Z","lastTransitionTime":"2025-11-25T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.327822 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.328045 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.328170 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.328258 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.328329 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:15Z","lastTransitionTime":"2025-11-25T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.430726 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.430970 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.431043 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.431123 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.431261 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:15Z","lastTransitionTime":"2025-11-25T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.504316 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:15 crc kubenswrapper[4802]: E1125 16:48:15.504438 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.534065 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.534096 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.534105 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.534141 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.534161 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:15Z","lastTransitionTime":"2025-11-25T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.636115 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.636177 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.636187 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.636202 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.636213 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:15Z","lastTransitionTime":"2025-11-25T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.738289 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.738332 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.738342 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.738359 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.738373 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:15Z","lastTransitionTime":"2025-11-25T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.841299 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.841349 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.841361 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.841378 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.841390 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:15Z","lastTransitionTime":"2025-11-25T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.944662 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.944733 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.944748 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.944773 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:15 crc kubenswrapper[4802]: I1125 16:48:15.944788 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:15Z","lastTransitionTime":"2025-11-25T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.047340 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.047397 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.047406 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.047423 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.047433 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:16Z","lastTransitionTime":"2025-11-25T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.150488 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.150525 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.150535 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.150550 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.150566 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:16Z","lastTransitionTime":"2025-11-25T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.253634 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.253691 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.253709 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.253731 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.253746 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:16Z","lastTransitionTime":"2025-11-25T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.356162 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.356231 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.356245 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.356262 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.356274 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:16Z","lastTransitionTime":"2025-11-25T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.458459 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.458509 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.458521 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.458540 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.458552 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:16Z","lastTransitionTime":"2025-11-25T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.504292 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.504341 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.504432 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:16 crc kubenswrapper[4802]: E1125 16:48:16.504452 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:16 crc kubenswrapper[4802]: E1125 16:48:16.504591 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:16 crc kubenswrapper[4802]: E1125 16:48:16.504664 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.565301 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.565384 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.565409 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.565425 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.565435 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:16Z","lastTransitionTime":"2025-11-25T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.668250 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.668290 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.668301 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.668318 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.668331 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:16Z","lastTransitionTime":"2025-11-25T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.770795 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.770832 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.770840 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.770853 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.770863 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:16Z","lastTransitionTime":"2025-11-25T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.872824 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.872869 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.872880 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.872895 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.872907 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:16Z","lastTransitionTime":"2025-11-25T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.975147 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.975189 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.975199 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.975215 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:16 crc kubenswrapper[4802]: I1125 16:48:16.975226 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:16Z","lastTransitionTime":"2025-11-25T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.078241 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.079193 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.079210 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.079228 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.079237 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:17Z","lastTransitionTime":"2025-11-25T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.182141 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.182191 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.182224 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.182243 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.182253 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:17Z","lastTransitionTime":"2025-11-25T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.284948 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.284991 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.285007 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.285023 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.285032 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:17Z","lastTransitionTime":"2025-11-25T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.387675 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.387706 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.387713 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.387727 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.387735 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:17Z","lastTransitionTime":"2025-11-25T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.490885 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.490948 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.490966 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.490989 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.491003 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:17Z","lastTransitionTime":"2025-11-25T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.503654 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:17 crc kubenswrapper[4802]: E1125 16:48:17.503837 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.520185 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.533619 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.545253 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.558352 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.571802 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.587042 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.592694 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.592725 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.592733 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.592748 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.592758 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:17Z","lastTransitionTime":"2025-11-25T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.603569 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.617032 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.628236 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0e948183d9361bf5f6831d25601431a1de0662987928a7277917451064c0295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:48:06Z\\\",\\\"message\\\":\\\"2025-11-25T16:47:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_07bab96e-be78-4d51-aab3-a524db7619cd\\\\n2025-11-25T16:47:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_07bab96e-be78-4d51-aab3-a524db7619cd to /host/opt/cni/bin/\\\\n2025-11-25T16:47:21Z [verbose] multus-daemon started\\\\n2025-11-25T16:47:21Z [verbose] Readiness Indicator file check\\\\n2025-11-25T16:48:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.638952 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.657890 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.671846 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.683897 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd9aa596-76c2-468c-b732-d85bc1f16070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6ft2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.694777 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.694815 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.694828 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.694846 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.694858 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:17Z","lastTransitionTime":"2025-11-25T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.696701 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feaf869b-f88a-4874-8174-b689fcf6f40a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d672b4cc7df9d1855c8e882c5f4c14dec0df01a8eadab587920753fee67abf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92fbcf119fd704d01bca613b32439360422fe8ea79ba3331eace6d7f52cb5d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7p6zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.710447 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fd65faa-292f-433b-83de-043c66cf484d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://060bdec5b0ff95b1ac4dcdb60f4bbb811ddeff0c93e74e373640bf65cb1d5ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb75f902fead2f9945afbecfdfd7b7fd35c5eaa4a05059a2038fb08e07875ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4bee24ceffdebad39410dfdca5f8a83c8664f6b4c18cc2495f47f47162d116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.731105 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88787834da7629acd6521f977ee73dec4231d0b4de6a3bb8f4dad83d203b4221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88787834da7629acd6521f977ee73dec4231d0b4de6a3bb8f4dad83d203b4221\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:48:11Z\\\",\\\"message\\\":\\\"column _uuid == {d8772e82-b0a4-4596-87d3-3d517c13344b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 16:48:11.274563 6791 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1125 16:48:11.274603 6791 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed call\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:48:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.740262 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.748641 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b4845-01e3-4ae3-9027-b6ffe66c4bc6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db167a028aae7e1be86605fcfb5a716b35c6e34d600b9675fb2c26b88bc52eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74e0f0d4a5a5b635060b11f8b506df6000fe73f589125d64bdc0450dec47ccbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74e0f0d4a5a5b635060b11f8b506df6000fe73f589125d64bdc0450dec47ccbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:17Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.796464 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.796505 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.796516 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.796533 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.796543 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:17Z","lastTransitionTime":"2025-11-25T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.900142 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.900238 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.900254 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.900287 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:17 crc kubenswrapper[4802]: I1125 16:48:17.900302 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:17Z","lastTransitionTime":"2025-11-25T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.003003 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.003072 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.003087 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.003106 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.003141 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:18Z","lastTransitionTime":"2025-11-25T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.105415 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.105492 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.105513 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.105536 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.105553 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:18Z","lastTransitionTime":"2025-11-25T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.208195 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.208241 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.208284 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.208298 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.208307 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:18Z","lastTransitionTime":"2025-11-25T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.310307 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.310347 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.310358 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.310372 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.310382 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:18Z","lastTransitionTime":"2025-11-25T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.412628 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.412686 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.412698 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.412717 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.412728 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:18Z","lastTransitionTime":"2025-11-25T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.503731 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.503882 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.503979 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:18 crc kubenswrapper[4802]: E1125 16:48:18.503905 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:18 crc kubenswrapper[4802]: E1125 16:48:18.504182 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:18 crc kubenswrapper[4802]: E1125 16:48:18.504243 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.515174 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.515235 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.515255 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.515284 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.515306 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:18Z","lastTransitionTime":"2025-11-25T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.617231 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.617262 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.617270 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.617284 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.617302 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:18Z","lastTransitionTime":"2025-11-25T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.719796 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.719832 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.719840 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.719854 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.719874 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:18Z","lastTransitionTime":"2025-11-25T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.822641 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.822709 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.822727 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.822757 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.822800 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:18Z","lastTransitionTime":"2025-11-25T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.925669 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.925713 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.925723 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.925737 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:18 crc kubenswrapper[4802]: I1125 16:48:18.925747 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:18Z","lastTransitionTime":"2025-11-25T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.028627 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.028691 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.028703 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.028721 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.028733 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:19Z","lastTransitionTime":"2025-11-25T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.131676 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.131724 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.131734 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.131754 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.131765 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:19Z","lastTransitionTime":"2025-11-25T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.233702 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.233764 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.233776 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.233789 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.233800 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:19Z","lastTransitionTime":"2025-11-25T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.336098 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.336184 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.336197 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.336214 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.336225 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:19Z","lastTransitionTime":"2025-11-25T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.438765 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.438795 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.438803 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.438818 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.438828 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:19Z","lastTransitionTime":"2025-11-25T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.504466 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:19 crc kubenswrapper[4802]: E1125 16:48:19.504642 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.540815 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.540865 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.540876 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.540893 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.540907 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:19Z","lastTransitionTime":"2025-11-25T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.644354 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.644475 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.644507 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.644547 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.644574 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:19Z","lastTransitionTime":"2025-11-25T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.748484 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.748552 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.748565 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.748592 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.748607 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:19Z","lastTransitionTime":"2025-11-25T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.851840 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.851906 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.851921 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.851942 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.851956 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:19Z","lastTransitionTime":"2025-11-25T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.955740 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.955800 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.955814 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.955835 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:19 crc kubenswrapper[4802]: I1125 16:48:19.955852 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:19Z","lastTransitionTime":"2025-11-25T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.059256 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.059344 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.059375 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.059411 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.059440 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:20Z","lastTransitionTime":"2025-11-25T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.162290 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.162348 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.162360 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.162380 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.162393 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:20Z","lastTransitionTime":"2025-11-25T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.265600 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.265707 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.265729 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.265761 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.265787 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:20Z","lastTransitionTime":"2025-11-25T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.369045 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.369342 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.369371 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.369409 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.369436 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:20Z","lastTransitionTime":"2025-11-25T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.474449 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.474533 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.474557 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.474592 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.474617 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:20Z","lastTransitionTime":"2025-11-25T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.504241 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.504284 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:20 crc kubenswrapper[4802]: E1125 16:48:20.504410 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.504320 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:20 crc kubenswrapper[4802]: E1125 16:48:20.504592 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:20 crc kubenswrapper[4802]: E1125 16:48:20.504699 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.534076 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:48:20 crc kubenswrapper[4802]: E1125 16:48:20.534280 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.534240761 +0000 UTC m=+147.678587977 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.534616 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.534758 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.534850 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:20 crc kubenswrapper[4802]: E1125 16:48:20.534857 4802 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 16:48:20 crc kubenswrapper[4802]: E1125 16:48:20.534934 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.53491913 +0000 UTC m=+147.679266546 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 16:48:20 crc kubenswrapper[4802]: E1125 16:48:20.535066 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 16:48:20 crc kubenswrapper[4802]: E1125 16:48:20.535106 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 16:48:20 crc kubenswrapper[4802]: E1125 16:48:20.535154 4802 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:48:20 crc kubenswrapper[4802]: E1125 16:48:20.535059 4802 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 16:48:20 crc kubenswrapper[4802]: E1125 16:48:20.535229 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.535207299 +0000 UTC m=+147.679554665 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:48:20 crc kubenswrapper[4802]: E1125 16:48:20.535307 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.535281231 +0000 UTC m=+147.679628417 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.577652 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.577698 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.577710 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.577727 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.577740 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:20Z","lastTransitionTime":"2025-11-25T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.635979 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:20 crc kubenswrapper[4802]: E1125 16:48:20.636205 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 16:48:20 crc kubenswrapper[4802]: E1125 16:48:20.636250 4802 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 16:48:20 crc kubenswrapper[4802]: E1125 16:48:20.636265 4802 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:48:20 crc kubenswrapper[4802]: E1125 16:48:20.636334 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.636315862 +0000 UTC m=+147.780663068 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.680565 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.680619 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.680634 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.680742 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.680774 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:20Z","lastTransitionTime":"2025-11-25T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.783004 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.783052 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.783062 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.783074 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.783084 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:20Z","lastTransitionTime":"2025-11-25T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.885616 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.885656 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.885666 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.885682 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.885692 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:20Z","lastTransitionTime":"2025-11-25T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.988212 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.988258 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.988278 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.988294 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:20 crc kubenswrapper[4802]: I1125 16:48:20.988302 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:20Z","lastTransitionTime":"2025-11-25T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.090573 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.090616 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.090628 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.090643 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.090651 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:21Z","lastTransitionTime":"2025-11-25T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.193279 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.193362 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.193381 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.193403 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.193421 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:21Z","lastTransitionTime":"2025-11-25T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.296535 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.296594 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.296606 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.296628 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.296652 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:21Z","lastTransitionTime":"2025-11-25T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.398638 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.398678 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.398692 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.398710 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.398722 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:21Z","lastTransitionTime":"2025-11-25T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.501450 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.501485 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.501494 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.501511 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.501522 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:21Z","lastTransitionTime":"2025-11-25T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.503995 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:21 crc kubenswrapper[4802]: E1125 16:48:21.504103 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.603925 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.603962 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.603972 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.603986 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.603997 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:21Z","lastTransitionTime":"2025-11-25T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.706262 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.706319 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.706335 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.706359 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.706375 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:21Z","lastTransitionTime":"2025-11-25T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.808706 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.808773 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.808790 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.808814 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.808831 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:21Z","lastTransitionTime":"2025-11-25T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.912273 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.912320 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.912332 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.912350 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:21 crc kubenswrapper[4802]: I1125 16:48:21.912362 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:21Z","lastTransitionTime":"2025-11-25T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.018260 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.018300 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.018310 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.018330 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.018416 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:22Z","lastTransitionTime":"2025-11-25T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.121229 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.121286 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.121297 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.121314 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.121325 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:22Z","lastTransitionTime":"2025-11-25T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.224594 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.224649 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.224658 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.224692 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.224702 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:22Z","lastTransitionTime":"2025-11-25T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.327999 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.328622 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.328641 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.328667 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.328686 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:22Z","lastTransitionTime":"2025-11-25T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.432826 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.432875 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.432884 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.432902 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.432914 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:22Z","lastTransitionTime":"2025-11-25T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.504418 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.504500 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.504453 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:22 crc kubenswrapper[4802]: E1125 16:48:22.504711 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:22 crc kubenswrapper[4802]: E1125 16:48:22.504862 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:22 crc kubenswrapper[4802]: E1125 16:48:22.504997 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.536706 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.536827 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.536853 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.536879 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.536899 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:22Z","lastTransitionTime":"2025-11-25T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.639795 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.639877 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.639901 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.639935 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.639961 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:22Z","lastTransitionTime":"2025-11-25T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.742304 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.742358 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.742369 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.742387 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.742398 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:22Z","lastTransitionTime":"2025-11-25T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.845195 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.845253 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.845271 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.845294 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.845312 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:22Z","lastTransitionTime":"2025-11-25T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.948619 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.948688 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.948705 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.948732 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:22 crc kubenswrapper[4802]: I1125 16:48:22.948750 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:22Z","lastTransitionTime":"2025-11-25T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.051993 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.052060 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.052083 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.052113 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.052167 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:23Z","lastTransitionTime":"2025-11-25T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.155623 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.155660 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.155668 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.155682 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.155696 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:23Z","lastTransitionTime":"2025-11-25T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.258544 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.258578 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.258586 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.258600 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.258610 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:23Z","lastTransitionTime":"2025-11-25T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.361065 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.361103 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.361140 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.361157 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.361168 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:23Z","lastTransitionTime":"2025-11-25T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.463267 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.463312 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.463326 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.463344 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.463355 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:23Z","lastTransitionTime":"2025-11-25T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.504487 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:23 crc kubenswrapper[4802]: E1125 16:48:23.504808 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.505051 4802 scope.go:117] "RemoveContainer" containerID="88787834da7629acd6521f977ee73dec4231d0b4de6a3bb8f4dad83d203b4221" Nov 25 16:48:23 crc kubenswrapper[4802]: E1125 16:48:23.505292 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.565944 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.565990 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.566017 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.566042 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.566058 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:23Z","lastTransitionTime":"2025-11-25T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.668077 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.668116 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.668148 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.668161 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.668169 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:23Z","lastTransitionTime":"2025-11-25T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.770354 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.770389 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.770401 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.770416 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.770429 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:23Z","lastTransitionTime":"2025-11-25T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.872229 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.872298 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.872322 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.872349 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.872367 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:23Z","lastTransitionTime":"2025-11-25T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.975043 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.975108 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.975165 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.975195 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:23 crc kubenswrapper[4802]: I1125 16:48:23.975217 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:23Z","lastTransitionTime":"2025-11-25T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.077611 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.077653 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.077666 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.077682 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.077693 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:24Z","lastTransitionTime":"2025-11-25T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.180257 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.180327 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.180343 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.180366 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.180386 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:24Z","lastTransitionTime":"2025-11-25T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.283487 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.283539 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.283552 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.283572 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.283588 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:24Z","lastTransitionTime":"2025-11-25T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.386177 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.386239 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.386256 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.386283 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.386300 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:24Z","lastTransitionTime":"2025-11-25T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.431628 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.431698 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.431715 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.431738 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.431756 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:24Z","lastTransitionTime":"2025-11-25T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:24 crc kubenswrapper[4802]: E1125 16:48:24.452577 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.457116 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.457168 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.457179 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.457194 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.457206 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:24Z","lastTransitionTime":"2025-11-25T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:24 crc kubenswrapper[4802]: E1125 16:48:24.476240 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.480578 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.480666 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.480686 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.480710 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.480728 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:24Z","lastTransitionTime":"2025-11-25T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:24 crc kubenswrapper[4802]: E1125 16:48:24.502631 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.503687 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.503724 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.503798 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:24 crc kubenswrapper[4802]: E1125 16:48:24.503863 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:24 crc kubenswrapper[4802]: E1125 16:48:24.503963 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:24 crc kubenswrapper[4802]: E1125 16:48:24.504097 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.506696 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.506743 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.506761 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.506784 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.506802 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:24Z","lastTransitionTime":"2025-11-25T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:24 crc kubenswrapper[4802]: E1125 16:48:24.526383 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.531655 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.531694 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.531705 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.531722 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.531735 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:24Z","lastTransitionTime":"2025-11-25T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:24 crc kubenswrapper[4802]: E1125 16:48:24.550885 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:24Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:24 crc kubenswrapper[4802]: E1125 16:48:24.551168 4802 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.553151 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.553189 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.553202 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.553219 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.553231 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:24Z","lastTransitionTime":"2025-11-25T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.656190 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.656261 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.656285 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.656313 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.656330 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:24Z","lastTransitionTime":"2025-11-25T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.759023 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.759084 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.759099 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.759116 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.759150 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:24Z","lastTransitionTime":"2025-11-25T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.861893 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.861956 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.861973 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.861996 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.862013 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:24Z","lastTransitionTime":"2025-11-25T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.964562 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.964684 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.964703 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.964725 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:24 crc kubenswrapper[4802]: I1125 16:48:24.964742 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:24Z","lastTransitionTime":"2025-11-25T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.067310 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.067361 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.067373 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.067392 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.067405 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:25Z","lastTransitionTime":"2025-11-25T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.170053 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.170089 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.170100 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.170114 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.170140 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:25Z","lastTransitionTime":"2025-11-25T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.272813 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.272864 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.272875 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.272895 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.272909 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:25Z","lastTransitionTime":"2025-11-25T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.375162 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.375218 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.375232 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.375255 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.375270 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:25Z","lastTransitionTime":"2025-11-25T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.477520 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.477548 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.477582 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.477599 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.477609 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:25Z","lastTransitionTime":"2025-11-25T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.504201 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:25 crc kubenswrapper[4802]: E1125 16:48:25.504307 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.579964 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.580012 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.580029 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.580051 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.580068 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:25Z","lastTransitionTime":"2025-11-25T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.682151 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.682215 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.682237 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.682268 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.682294 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:25Z","lastTransitionTime":"2025-11-25T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.784590 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.784634 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.784646 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.784661 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.784675 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:25Z","lastTransitionTime":"2025-11-25T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.886426 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.886496 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.886508 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.886522 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.886530 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:25Z","lastTransitionTime":"2025-11-25T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.988612 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.988655 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.988667 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.988683 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:25 crc kubenswrapper[4802]: I1125 16:48:25.988695 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:25Z","lastTransitionTime":"2025-11-25T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.090600 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.090639 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.090647 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.090660 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.090672 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:26Z","lastTransitionTime":"2025-11-25T16:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.193362 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.193405 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.193415 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.193432 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.193441 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:26Z","lastTransitionTime":"2025-11-25T16:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.295430 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.295463 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.295475 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.295489 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.295497 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:26Z","lastTransitionTime":"2025-11-25T16:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.400563 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.400704 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.400727 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.400747 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.400759 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:26Z","lastTransitionTime":"2025-11-25T16:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.503214 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.503265 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.503276 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.503293 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.503305 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:26Z","lastTransitionTime":"2025-11-25T16:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.503472 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.503509 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.503564 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:26 crc kubenswrapper[4802]: E1125 16:48:26.503565 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:26 crc kubenswrapper[4802]: E1125 16:48:26.503648 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:26 crc kubenswrapper[4802]: E1125 16:48:26.503718 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.605904 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.605943 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.605952 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.605968 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.605980 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:26Z","lastTransitionTime":"2025-11-25T16:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.708421 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.708460 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.708469 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.708485 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.708494 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:26Z","lastTransitionTime":"2025-11-25T16:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.811495 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.811780 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.811877 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.811986 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.812084 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:26Z","lastTransitionTime":"2025-11-25T16:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.915054 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.915105 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.915142 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.915165 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:26 crc kubenswrapper[4802]: I1125 16:48:26.915178 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:26Z","lastTransitionTime":"2025-11-25T16:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.018075 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.018262 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.018278 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.018296 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.018308 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:27Z","lastTransitionTime":"2025-11-25T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.120886 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.120971 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.120982 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.120998 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.121008 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:27Z","lastTransitionTime":"2025-11-25T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.223152 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.223222 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.223230 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.223245 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.223254 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:27Z","lastTransitionTime":"2025-11-25T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.326532 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.326586 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.326599 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.326617 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.326630 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:27Z","lastTransitionTime":"2025-11-25T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.429292 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.429357 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.429370 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.429384 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.429393 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:27Z","lastTransitionTime":"2025-11-25T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.504542 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:27 crc kubenswrapper[4802]: E1125 16:48:27.504684 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.517937 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.530140 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.531568 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.531606 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.531615 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.531631 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.531641 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:27Z","lastTransitionTime":"2025-11-25T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.542412 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471630afa22e7ad244bf7c5fa8b8f0722ae8bcd56f561ac82003c04a56ea14d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f3ef920b7d88197d8d9167768aa154f318e200e4d41426b904ef9f8351f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.552275 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dqg6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7677a2dc-6b77-4fbc-90e3-9548a2298016\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9115fb985068563862d632892db0b0024341aaf5344805907e05444e164c2fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ghr7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dqg6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.563463 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9435c266-8324-4e76-bdc6-a0a63ab94230\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.576780 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e31b74d-66a6-4a22-a22d-ed8233064c19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722a092dd07ead22c345cf30d0b70c3b0c557c6431572b7b060c6e345aeb700e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32f0d28432423527d4d8405b7a7290f26335f0d30915ce03322be1eb62e5044e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3c88f6eec6ba668bfcb67cba7655ab637c30a67f3ae91e6fbb7a3a7270f92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.587993 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f7b0f07e0c25173db1acb6d65cd4a1aa12f1dc56f6a7cb0178460b190647611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.605806 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.618397 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63624d38dcbfbc019fb65e83275665d57ed16cb74d73261ecfac920d22e57097\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.629585 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hmbvn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97e822d6-58fe-41f4-b08b-3c9b42273307\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0e948183d9361bf5f6831d25601431a1de0662987928a7277917451064c0295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:48:06Z\\\",\\\"message\\\":\\\"2025-11-25T16:47:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_07bab96e-be78-4d51-aab3-a524db7619cd\\\\n2025-11-25T16:47:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_07bab96e-be78-4d51-aab3-a524db7619cd to /host/opt/cni/bin/\\\\n2025-11-25T16:47:21Z [verbose] multus-daemon started\\\\n2025-11-25T16:47:21Z [verbose] Readiness Indicator file check\\\\n2025-11-25T16:48:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:48:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md8dh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hmbvn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.633391 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.633432 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.633451 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.633466 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.633475 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:27Z","lastTransitionTime":"2025-11-25T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.640781 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5958b72b628c31856c35db045e8b0de0094df2b8f3461d673170a62f4ea23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrbk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h29wc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.653996 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qclgb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32628afd-a82e-48a8-a3d2-cfb23d5ba37b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cc2a9d5751365a6515dae6d8eb737a1e1921998ab16e3c5aade6aaeb0c80862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9259e14e05c98a63ce474ff9140035363020b7c08a4277fe30071e9e322e843f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a914fc700ced26abf1be062dd3e74185cc5065142d7276d569674f06ff10388a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8f0b1f7c72aedb2b3ffa6df897088035cc3a3d900c55952cea089b87e5c0bcb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c89395e747afd1fdbbf81e6cf75daef01aeffd85b4175ff3112cc0e7b235ca94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7688591ac85dceb0435c35ee1c09a55441fecabbec08591fa3e8946bad4db6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bb2dfb3dcf447fd76f66d90c017fc2c16e3a90567f1bccf895cc811a92bc7b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xc62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qclgb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.664742 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd9aa596-76c2-468c-b732-d85bc1f16070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6ft2z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.676435 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fd65faa-292f-433b-83de-043c66cf484d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://060bdec5b0ff95b1ac4dcdb60f4bbb811ddeff0c93e74e373640bf65cb1d5ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ceb75f902fead2f9945afbecfdfd7b7fd35c5eaa4a05059a2038fb08e07875ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4bee24ceffdebad39410dfdca5f8a83c8664f6b4c18cc2495f47f47162d116a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44915749b26fdd592bd044f24e191a41e20b952a92dda760c34baba3d4b2cd45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.686854 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"feaf869b-f88a-4874-8174-b689fcf6f40a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d672b4cc7df9d1855c8e882c5f4c14dec0df01a8eadab587920753fee67abf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92fbcf119fd704d01bca613b32439360422fe8ea79ba3331eace6d7f52cb5d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tqsfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7p6zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.697575 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b4845-01e3-4ae3-9027-b6ffe66c4bc6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db167a028aae7e1be86605fcfb5a716b35c6e34d600b9675fb2c26b88bc52eb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74e0f0d4a5a5b635060b11f8b506df6000fe73f589125d64bdc0450dec47ccbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74e0f0d4a5a5b635060b11f8b506df6000fe73f589125d64bdc0450dec47ccbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:46:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:46:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.714272 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad5c073-f5d9-4410-9350-bb2a51a764a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88787834da7629acd6521f977ee73dec4231d0b4de6a3bb8f4dad83d203b4221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88787834da7629acd6521f977ee73dec4231d0b4de6a3bb8f4dad83d203b4221\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T16:48:11Z\\\",\\\"message\\\":\\\"column _uuid == {d8772e82-b0a4-4596-87d3-3d517c13344b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 16:48:11.274563 6791 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1125 16:48:11.274603 6791 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed call\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T16:48:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T16:47:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T16:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgb5h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dkxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.723788 4802 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pbmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f6394dc-d7bd-4205-9ff4-aca84d1306b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T16:47:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bed0ba35218419c785d10df252e5805cfc31f19f7ae48829d6ffd27b87adde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T16:47:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kn92s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T16:47:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pbmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:27Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.735459 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.735501 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.735510 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.735524 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.735536 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:27Z","lastTransitionTime":"2025-11-25T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.837912 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.837975 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.837985 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.837999 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.838008 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:27Z","lastTransitionTime":"2025-11-25T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.940205 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.940242 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.940251 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.940267 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:27 crc kubenswrapper[4802]: I1125 16:48:27.940275 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:27Z","lastTransitionTime":"2025-11-25T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.043352 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.043390 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.043398 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.043409 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.043418 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:28Z","lastTransitionTime":"2025-11-25T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.146179 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.146245 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.146263 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.146291 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.146363 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:28Z","lastTransitionTime":"2025-11-25T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.249151 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.249200 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.249212 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.249231 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.249242 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:28Z","lastTransitionTime":"2025-11-25T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.351630 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.351675 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.351687 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.351711 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.351726 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:28Z","lastTransitionTime":"2025-11-25T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.454153 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.454210 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.454224 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.454242 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.454257 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:28Z","lastTransitionTime":"2025-11-25T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.503465 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.503536 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:28 crc kubenswrapper[4802]: E1125 16:48:28.503577 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.503613 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:28 crc kubenswrapper[4802]: E1125 16:48:28.503687 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:28 crc kubenswrapper[4802]: E1125 16:48:28.503757 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.557060 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.557155 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.557173 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.557196 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.557213 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:28Z","lastTransitionTime":"2025-11-25T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.660117 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.660201 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.660211 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.660232 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.660244 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:28Z","lastTransitionTime":"2025-11-25T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.763467 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.763540 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.763552 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.763571 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.763906 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:28Z","lastTransitionTime":"2025-11-25T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.866215 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.866281 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.866296 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.866317 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.866332 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:28Z","lastTransitionTime":"2025-11-25T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.968719 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.968762 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.968771 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.968785 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:28 crc kubenswrapper[4802]: I1125 16:48:28.968794 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:28Z","lastTransitionTime":"2025-11-25T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.071870 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.071952 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.071970 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.072006 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.072029 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:29Z","lastTransitionTime":"2025-11-25T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.174952 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.174987 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.174999 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.175015 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.175025 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:29Z","lastTransitionTime":"2025-11-25T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.278093 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.278204 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.278274 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.278303 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.278326 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:29Z","lastTransitionTime":"2025-11-25T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.380220 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.380249 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.380259 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.380273 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.380291 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:29Z","lastTransitionTime":"2025-11-25T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.482427 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.482486 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.482498 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.482514 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.482525 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:29Z","lastTransitionTime":"2025-11-25T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.503824 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:29 crc kubenswrapper[4802]: E1125 16:48:29.504067 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.585725 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.585775 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.585786 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.585805 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.585816 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:29Z","lastTransitionTime":"2025-11-25T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.689110 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.689181 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.689194 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.689214 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.689225 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:29Z","lastTransitionTime":"2025-11-25T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.791613 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.791657 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.791667 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.791681 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.791693 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:29Z","lastTransitionTime":"2025-11-25T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.894352 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.894422 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.894431 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.894448 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.894458 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:29Z","lastTransitionTime":"2025-11-25T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.997071 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.997150 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.997161 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.997176 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:29 crc kubenswrapper[4802]: I1125 16:48:29.997188 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:29Z","lastTransitionTime":"2025-11-25T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.099266 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.099319 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.099330 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.099342 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.099351 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:30Z","lastTransitionTime":"2025-11-25T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.204760 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.204832 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.204852 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.204876 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.204900 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:30Z","lastTransitionTime":"2025-11-25T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.308275 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.308343 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.308361 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.308388 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.308427 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:30Z","lastTransitionTime":"2025-11-25T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.411465 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.411525 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.411544 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.411568 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.411584 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:30Z","lastTransitionTime":"2025-11-25T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.503646 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.503713 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.503988 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:30 crc kubenswrapper[4802]: E1125 16:48:30.504070 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:30 crc kubenswrapper[4802]: E1125 16:48:30.504245 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:30 crc kubenswrapper[4802]: E1125 16:48:30.504283 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.513999 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.514068 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.514092 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.514152 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.514180 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:30Z","lastTransitionTime":"2025-11-25T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.617066 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.617105 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.617113 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.617143 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.617152 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:30Z","lastTransitionTime":"2025-11-25T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.720489 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.720551 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.720575 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.720602 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.720624 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:30Z","lastTransitionTime":"2025-11-25T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.826056 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.826112 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.826150 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.826167 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.826177 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:30Z","lastTransitionTime":"2025-11-25T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.928685 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.928731 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.928767 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.928787 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:30 crc kubenswrapper[4802]: I1125 16:48:30.928800 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:30Z","lastTransitionTime":"2025-11-25T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.031798 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.031837 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.031849 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.031866 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.031878 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:31Z","lastTransitionTime":"2025-11-25T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.134627 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.134672 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.134681 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.134696 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.134704 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:31Z","lastTransitionTime":"2025-11-25T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.237339 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.237378 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.237390 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.237405 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.237416 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:31Z","lastTransitionTime":"2025-11-25T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.340162 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.340204 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.340216 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.340235 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.340246 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:31Z","lastTransitionTime":"2025-11-25T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.442948 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.442986 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.443007 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.443024 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.443038 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:31Z","lastTransitionTime":"2025-11-25T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.504531 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:31 crc kubenswrapper[4802]: E1125 16:48:31.504676 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.545438 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.545479 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.545489 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.545503 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.545514 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:31Z","lastTransitionTime":"2025-11-25T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.647950 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.647998 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.648010 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.648027 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.648038 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:31Z","lastTransitionTime":"2025-11-25T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.750188 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.750252 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.750263 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.750275 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.750285 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:31Z","lastTransitionTime":"2025-11-25T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.851929 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.851975 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.851985 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.852000 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.852011 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:31Z","lastTransitionTime":"2025-11-25T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.954462 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.954507 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.954541 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.954559 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:31 crc kubenswrapper[4802]: I1125 16:48:31.954570 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:31Z","lastTransitionTime":"2025-11-25T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.056942 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.056980 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.056996 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.057019 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.057034 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:32Z","lastTransitionTime":"2025-11-25T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.159658 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.159687 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.159695 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.159709 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.159718 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:32Z","lastTransitionTime":"2025-11-25T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.261814 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.261851 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.261859 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.261873 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.261882 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:32Z","lastTransitionTime":"2025-11-25T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.364590 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.364645 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.364662 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.364679 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.364690 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:32Z","lastTransitionTime":"2025-11-25T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.466473 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.466509 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.466517 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.466531 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.466541 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:32Z","lastTransitionTime":"2025-11-25T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.504006 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.504053 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.504027 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:32 crc kubenswrapper[4802]: E1125 16:48:32.504182 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:32 crc kubenswrapper[4802]: E1125 16:48:32.504279 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:32 crc kubenswrapper[4802]: E1125 16:48:32.504363 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.568586 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.568643 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.568652 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.568669 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.568678 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:32Z","lastTransitionTime":"2025-11-25T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.671033 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.671073 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.671084 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.671098 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.671109 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:32Z","lastTransitionTime":"2025-11-25T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.774514 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.774543 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.774551 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.774563 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.774572 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:32Z","lastTransitionTime":"2025-11-25T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.876456 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.876483 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.876491 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.876503 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.876512 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:32Z","lastTransitionTime":"2025-11-25T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.982026 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.982077 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.982087 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.982104 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:32 crc kubenswrapper[4802]: I1125 16:48:32.982114 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:32Z","lastTransitionTime":"2025-11-25T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.084292 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.084325 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.084333 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.084351 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.084363 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:33Z","lastTransitionTime":"2025-11-25T16:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.186182 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.186216 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.186225 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.186241 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.186251 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:33Z","lastTransitionTime":"2025-11-25T16:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.288369 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.288616 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.288693 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.288766 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.288839 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:33Z","lastTransitionTime":"2025-11-25T16:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.390804 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.390840 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.390852 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.390866 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.390876 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:33Z","lastTransitionTime":"2025-11-25T16:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.493102 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.493157 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.493167 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.493181 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.493190 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:33Z","lastTransitionTime":"2025-11-25T16:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.503876 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:33 crc kubenswrapper[4802]: E1125 16:48:33.503996 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.595084 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.595908 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.596001 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.596088 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.596194 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:33Z","lastTransitionTime":"2025-11-25T16:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.698216 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.698479 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.698573 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.698684 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.698749 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:33Z","lastTransitionTime":"2025-11-25T16:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.801437 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.801483 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.801499 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.801516 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.801527 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:33Z","lastTransitionTime":"2025-11-25T16:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.903747 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.903787 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.903795 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.903808 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:33 crc kubenswrapper[4802]: I1125 16:48:33.903816 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:33Z","lastTransitionTime":"2025-11-25T16:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.006139 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.006179 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.006195 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.006212 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.006223 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:34Z","lastTransitionTime":"2025-11-25T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.107527 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.107554 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.107564 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.107579 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.107591 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:34Z","lastTransitionTime":"2025-11-25T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.209680 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.209712 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.209720 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.209733 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.209742 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:34Z","lastTransitionTime":"2025-11-25T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.311029 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.311070 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.311081 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.311097 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.311111 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:34Z","lastTransitionTime":"2025-11-25T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.414033 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.414115 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.414152 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.414178 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.414198 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:34Z","lastTransitionTime":"2025-11-25T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.504416 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:34 crc kubenswrapper[4802]: E1125 16:48:34.504557 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.504448 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:34 crc kubenswrapper[4802]: E1125 16:48:34.504624 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.504427 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:34 crc kubenswrapper[4802]: E1125 16:48:34.505006 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.517414 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.517450 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.517464 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.517483 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.517499 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:34Z","lastTransitionTime":"2025-11-25T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.578805 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.578852 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.578862 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.578881 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.578895 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:34Z","lastTransitionTime":"2025-11-25T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:34 crc kubenswrapper[4802]: E1125 16:48:34.589894 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:34Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.592932 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.592965 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.592974 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.592989 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.593001 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:34Z","lastTransitionTime":"2025-11-25T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:34 crc kubenswrapper[4802]: E1125 16:48:34.605001 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:34Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.608253 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.608408 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.608436 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.608465 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.608487 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:34Z","lastTransitionTime":"2025-11-25T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:34 crc kubenswrapper[4802]: E1125 16:48:34.620184 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:34Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.623462 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.623502 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.623513 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.623533 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.623546 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:34Z","lastTransitionTime":"2025-11-25T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:34 crc kubenswrapper[4802]: E1125 16:48:34.638700 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:34Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.641481 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.641514 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.641525 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.641544 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.641557 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:34Z","lastTransitionTime":"2025-11-25T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:34 crc kubenswrapper[4802]: E1125 16:48:34.651949 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T16:48:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"190f8f70-36fb-445d-a11a-52e80e3afc28\\\",\\\"systemUUID\\\":\\\"7331689e-1a93-42fa-8eca-f0913b387d6b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T16:48:34Z is after 2025-08-24T17:21:41Z" Nov 25 16:48:34 crc kubenswrapper[4802]: E1125 16:48:34.652107 4802 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.653435 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.653464 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.653474 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.653488 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.653497 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:34Z","lastTransitionTime":"2025-11-25T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.755852 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.755896 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.755910 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.755927 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.755940 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:34Z","lastTransitionTime":"2025-11-25T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.858492 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.858533 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.858541 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.858555 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.858564 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:34Z","lastTransitionTime":"2025-11-25T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.960990 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.961031 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.961041 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.961056 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:34 crc kubenswrapper[4802]: I1125 16:48:34.961065 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:34Z","lastTransitionTime":"2025-11-25T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.063899 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.064269 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.064280 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.064294 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.064305 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:35Z","lastTransitionTime":"2025-11-25T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.166582 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.166647 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.166660 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.166682 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.166696 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:35Z","lastTransitionTime":"2025-11-25T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.269419 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.269466 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.269478 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.269497 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.269509 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:35Z","lastTransitionTime":"2025-11-25T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.372250 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.372289 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.372299 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.372316 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.372326 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:35Z","lastTransitionTime":"2025-11-25T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.474909 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.474945 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.474953 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.474967 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.474976 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:35Z","lastTransitionTime":"2025-11-25T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.504370 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:35 crc kubenswrapper[4802]: E1125 16:48:35.504481 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.577009 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.577061 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.577079 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.577103 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.577158 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:35Z","lastTransitionTime":"2025-11-25T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.680108 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.680190 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.680230 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.680249 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.680264 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:35Z","lastTransitionTime":"2025-11-25T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.782672 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.782712 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.782724 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.782739 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.782750 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:35Z","lastTransitionTime":"2025-11-25T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.884982 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.885011 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.885035 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.885050 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.885059 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:35Z","lastTransitionTime":"2025-11-25T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.987554 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.987626 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.987637 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.987653 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:35 crc kubenswrapper[4802]: I1125 16:48:35.987664 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:35Z","lastTransitionTime":"2025-11-25T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.089768 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.089797 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.089806 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.089818 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.089827 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:36Z","lastTransitionTime":"2025-11-25T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.191926 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.191961 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.191969 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.191982 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.191993 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:36Z","lastTransitionTime":"2025-11-25T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.293841 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.293885 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.293902 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.293920 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.293933 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:36Z","lastTransitionTime":"2025-11-25T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.396547 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.396604 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.396616 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.396632 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.396642 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:36Z","lastTransitionTime":"2025-11-25T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.498824 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.498860 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.498868 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.498881 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.498889 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:36Z","lastTransitionTime":"2025-11-25T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.504251 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.504316 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.504347 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:36 crc kubenswrapper[4802]: E1125 16:48:36.504401 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:36 crc kubenswrapper[4802]: E1125 16:48:36.504983 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:36 crc kubenswrapper[4802]: E1125 16:48:36.505156 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.515991 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.594989 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs\") pod \"network-metrics-daemon-6ft2z\" (UID: \"cd9aa596-76c2-468c-b732-d85bc1f16070\") " pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:36 crc kubenswrapper[4802]: E1125 16:48:36.595625 4802 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 16:48:36 crc kubenswrapper[4802]: E1125 16:48:36.595727 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs podName:cd9aa596-76c2-468c-b732-d85bc1f16070 nodeName:}" failed. No retries permitted until 2025-11-25 16:49:40.595707467 +0000 UTC m=+163.740054653 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs") pod "network-metrics-daemon-6ft2z" (UID: "cd9aa596-76c2-468c-b732-d85bc1f16070") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.601317 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.601345 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.601355 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.601369 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.601380 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:36Z","lastTransitionTime":"2025-11-25T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.703629 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.703658 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.703666 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.703680 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.703689 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:36Z","lastTransitionTime":"2025-11-25T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.805737 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.805776 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.805786 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.805803 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.805815 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:36Z","lastTransitionTime":"2025-11-25T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.908167 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.908242 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.908251 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.908288 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:36 crc kubenswrapper[4802]: I1125 16:48:36.908298 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:36Z","lastTransitionTime":"2025-11-25T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.012249 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.012349 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.012367 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.012394 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.012408 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:37Z","lastTransitionTime":"2025-11-25T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.115254 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.115510 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.115632 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.115820 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.115958 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:37Z","lastTransitionTime":"2025-11-25T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.218449 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.218507 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.218526 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.218549 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.218565 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:37Z","lastTransitionTime":"2025-11-25T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.321149 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.321200 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.321212 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.321228 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.321239 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:37Z","lastTransitionTime":"2025-11-25T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.423748 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.424003 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.424077 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.424175 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.424242 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:37Z","lastTransitionTime":"2025-11-25T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.504300 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:37 crc kubenswrapper[4802]: E1125 16:48:37.504442 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.523730 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=25.523711255 podStartE2EDuration="25.523711255s" podCreationTimestamp="2025-11-25 16:48:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:48:37.523696485 +0000 UTC m=+100.668043701" watchObservedRunningTime="2025-11-25 16:48:37.523711255 +0000 UTC m=+100.668058441" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.526317 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.526390 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.526415 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.526444 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.526467 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:37Z","lastTransitionTime":"2025-11-25T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.559209 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=1.559185378 podStartE2EDuration="1.559185378s" podCreationTimestamp="2025-11-25 16:48:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:48:37.557831733 +0000 UTC m=+100.702178959" watchObservedRunningTime="2025-11-25 16:48:37.559185378 +0000 UTC m=+100.703532594" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.599275 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-2pbmn" podStartSLOduration=80.599258641 podStartE2EDuration="1m20.599258641s" podCreationTimestamp="2025-11-25 16:47:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:48:37.598617464 +0000 UTC m=+100.742964700" watchObservedRunningTime="2025-11-25 16:48:37.599258641 +0000 UTC m=+100.743605827" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.628712 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.628735 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.628743 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.628757 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.628765 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:37Z","lastTransitionTime":"2025-11-25T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.648274 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-dqg6t" podStartSLOduration=80.648251516 podStartE2EDuration="1m20.648251516s" podCreationTimestamp="2025-11-25 16:47:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:48:37.64798645 +0000 UTC m=+100.792333646" watchObservedRunningTime="2025-11-25 16:48:37.648251516 +0000 UTC m=+100.792598702" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.679220 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=81.679202652 podStartE2EDuration="1m21.679202652s" podCreationTimestamp="2025-11-25 16:47:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:48:37.66644887 +0000 UTC m=+100.810796076" watchObservedRunningTime="2025-11-25 16:48:37.679202652 +0000 UTC m=+100.823549828" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.690279 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=74.69026446 podStartE2EDuration="1m14.69026446s" podCreationTimestamp="2025-11-25 16:47:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:48:37.679004317 +0000 UTC m=+100.823351503" watchObservedRunningTime="2025-11-25 16:48:37.69026446 +0000 UTC m=+100.834611646" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.728964 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-hmbvn" podStartSLOduration=79.728946296 podStartE2EDuration="1m19.728946296s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:48:37.72873729 +0000 UTC m=+100.873084496" watchObservedRunningTime="2025-11-25 16:48:37.728946296 +0000 UTC m=+100.873293482" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.730481 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.730517 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.730529 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.730546 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.730558 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:37Z","lastTransitionTime":"2025-11-25T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.745718 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podStartSLOduration=79.745699851 podStartE2EDuration="1m19.745699851s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:48:37.745383953 +0000 UTC m=+100.889731139" watchObservedRunningTime="2025-11-25 16:48:37.745699851 +0000 UTC m=+100.890047037" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.759628 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-qclgb" podStartSLOduration=79.759602543 podStartE2EDuration="1m19.759602543s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:48:37.758816683 +0000 UTC m=+100.903163859" watchObservedRunningTime="2025-11-25 16:48:37.759602543 +0000 UTC m=+100.903949729" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.781523 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=52.781506274 podStartE2EDuration="52.781506274s" podCreationTimestamp="2025-11-25 16:47:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:48:37.781057352 +0000 UTC m=+100.925404558" watchObservedRunningTime="2025-11-25 16:48:37.781506274 +0000 UTC m=+100.925853460" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.792427 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7p6zh" podStartSLOduration=79.792410488 podStartE2EDuration="1m19.792410488s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:48:37.792315065 +0000 UTC m=+100.936662271" watchObservedRunningTime="2025-11-25 16:48:37.792410488 +0000 UTC m=+100.936757674" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.833404 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.833444 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.833455 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.833474 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.833486 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:37Z","lastTransitionTime":"2025-11-25T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.934984 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.935029 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.935038 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.935057 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:37 crc kubenswrapper[4802]: I1125 16:48:37.935066 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:37Z","lastTransitionTime":"2025-11-25T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.037259 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.037297 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.037305 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.037318 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.037326 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:38Z","lastTransitionTime":"2025-11-25T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.139081 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.139172 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.139185 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.139203 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.139215 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:38Z","lastTransitionTime":"2025-11-25T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.241751 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.241829 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.241850 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.241872 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.241887 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:38Z","lastTransitionTime":"2025-11-25T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.344381 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.344442 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.344460 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.344480 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.344495 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:38Z","lastTransitionTime":"2025-11-25T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.447054 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.447084 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.447094 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.447107 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.447117 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:38Z","lastTransitionTime":"2025-11-25T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.503583 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.503627 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.503607 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:38 crc kubenswrapper[4802]: E1125 16:48:38.503736 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:38 crc kubenswrapper[4802]: E1125 16:48:38.503863 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:38 crc kubenswrapper[4802]: E1125 16:48:38.503921 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.504461 4802 scope.go:117] "RemoveContainer" containerID="88787834da7629acd6521f977ee73dec4231d0b4de6a3bb8f4dad83d203b4221" Nov 25 16:48:38 crc kubenswrapper[4802]: E1125 16:48:38.504607 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dkxhj_openshift-ovn-kubernetes(bad5c073-f5d9-4410-9350-bb2a51a764a2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.550026 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.550085 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.550102 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.550157 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.550177 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:38Z","lastTransitionTime":"2025-11-25T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.653595 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.653904 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.654138 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.654236 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.654323 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:38Z","lastTransitionTime":"2025-11-25T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.756912 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.757485 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.757567 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.757691 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.757769 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:38Z","lastTransitionTime":"2025-11-25T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.860547 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.860587 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.860597 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.860612 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.860623 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:38Z","lastTransitionTime":"2025-11-25T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.963332 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.963368 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.963378 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.963397 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:38 crc kubenswrapper[4802]: I1125 16:48:38.963408 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:38Z","lastTransitionTime":"2025-11-25T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.066209 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.066250 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.066259 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.066274 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.066284 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:39Z","lastTransitionTime":"2025-11-25T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.168861 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.168941 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.168960 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.168992 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.169016 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:39Z","lastTransitionTime":"2025-11-25T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.271854 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.271939 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.271958 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.271983 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.272000 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:39Z","lastTransitionTime":"2025-11-25T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.374934 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.374980 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.374988 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.375004 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.375014 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:39Z","lastTransitionTime":"2025-11-25T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.477238 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.477299 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.477316 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.477340 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.477360 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:39Z","lastTransitionTime":"2025-11-25T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.504640 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:39 crc kubenswrapper[4802]: E1125 16:48:39.504972 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.579892 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.579941 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.579951 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.579969 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.579981 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:39Z","lastTransitionTime":"2025-11-25T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.682664 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.682700 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.682708 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.682722 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.682731 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:39Z","lastTransitionTime":"2025-11-25T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.784957 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.784989 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.785005 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.785020 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.785030 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:39Z","lastTransitionTime":"2025-11-25T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.888380 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.888413 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.888425 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.888442 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.888454 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:39Z","lastTransitionTime":"2025-11-25T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.990454 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.990525 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.990542 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.990564 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:39 crc kubenswrapper[4802]: I1125 16:48:39.990581 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:39Z","lastTransitionTime":"2025-11-25T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.092876 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.092916 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.092959 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.092995 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.093009 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:40Z","lastTransitionTime":"2025-11-25T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.195025 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.195170 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.195220 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.195246 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.195264 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:40Z","lastTransitionTime":"2025-11-25T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.298740 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.298889 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.298936 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.298972 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.298993 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:40Z","lastTransitionTime":"2025-11-25T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.401685 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.401785 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.401815 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.401842 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.401874 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:40Z","lastTransitionTime":"2025-11-25T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.503885 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.504035 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.504083 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.504155 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.504168 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.504184 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.504195 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:40Z","lastTransitionTime":"2025-11-25T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.504274 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:40 crc kubenswrapper[4802]: E1125 16:48:40.504393 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:40 crc kubenswrapper[4802]: E1125 16:48:40.504451 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:40 crc kubenswrapper[4802]: E1125 16:48:40.504509 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.606326 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.606388 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.606400 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.606414 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.606423 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:40Z","lastTransitionTime":"2025-11-25T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.710254 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.710319 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.710337 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.710367 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.710388 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:40Z","lastTransitionTime":"2025-11-25T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.813305 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.813365 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.813414 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.813439 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.813457 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:40Z","lastTransitionTime":"2025-11-25T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.916497 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.916551 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.916563 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.916583 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:40 crc kubenswrapper[4802]: I1125 16:48:40.916595 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:40Z","lastTransitionTime":"2025-11-25T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.020420 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.020505 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.020527 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.020558 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.020581 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:41Z","lastTransitionTime":"2025-11-25T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.123644 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.123708 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.123742 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.123774 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.123795 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:41Z","lastTransitionTime":"2025-11-25T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.227176 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.227256 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.227281 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.227310 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.227332 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:41Z","lastTransitionTime":"2025-11-25T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.330213 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.330278 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.330292 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.330314 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.330335 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:41Z","lastTransitionTime":"2025-11-25T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.434930 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.435005 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.435024 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.435053 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.435072 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:41Z","lastTransitionTime":"2025-11-25T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.504049 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:41 crc kubenswrapper[4802]: E1125 16:48:41.504399 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.537641 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.537696 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.537711 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.537735 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.537747 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:41Z","lastTransitionTime":"2025-11-25T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.640340 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.640418 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.640435 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.640456 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.640472 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:41Z","lastTransitionTime":"2025-11-25T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.742585 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.742635 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.742650 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.742667 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.742679 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:41Z","lastTransitionTime":"2025-11-25T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.844892 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.844929 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.844938 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.844952 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.844962 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:41Z","lastTransitionTime":"2025-11-25T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.947028 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.947061 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.947070 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.947134 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:41 crc kubenswrapper[4802]: I1125 16:48:41.947167 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:41Z","lastTransitionTime":"2025-11-25T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.049712 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.049746 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.049754 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.049767 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.049775 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:42Z","lastTransitionTime":"2025-11-25T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.151606 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.151646 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.151657 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.151687 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.151697 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:42Z","lastTransitionTime":"2025-11-25T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.253433 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.253464 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.253473 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.253519 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.253527 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:42Z","lastTransitionTime":"2025-11-25T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.355786 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.355823 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.355834 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.355849 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.355860 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:42Z","lastTransitionTime":"2025-11-25T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.457995 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.458036 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.458053 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.458068 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.458077 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:42Z","lastTransitionTime":"2025-11-25T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.503932 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.503932 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:42 crc kubenswrapper[4802]: E1125 16:48:42.504051 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.504229 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:42 crc kubenswrapper[4802]: E1125 16:48:42.504361 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:42 crc kubenswrapper[4802]: E1125 16:48:42.504473 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.559884 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.559921 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.559930 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.559945 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.559958 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:42Z","lastTransitionTime":"2025-11-25T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.662050 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.662298 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.662314 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.662329 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.662340 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:42Z","lastTransitionTime":"2025-11-25T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.764315 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.764350 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.764359 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.764371 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.764380 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:42Z","lastTransitionTime":"2025-11-25T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.866785 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.866817 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.866827 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.866840 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.866902 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:42Z","lastTransitionTime":"2025-11-25T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.969365 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.969419 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.969431 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.969450 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:42 crc kubenswrapper[4802]: I1125 16:48:42.969463 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:42Z","lastTransitionTime":"2025-11-25T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.071534 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.071579 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.071590 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.071606 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.071616 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:43Z","lastTransitionTime":"2025-11-25T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.174004 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.174113 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.174181 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.174213 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.174227 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:43Z","lastTransitionTime":"2025-11-25T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.275750 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.275784 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.275792 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.275806 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.275814 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:43Z","lastTransitionTime":"2025-11-25T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.379169 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.379214 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.379230 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.379256 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.379269 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:43Z","lastTransitionTime":"2025-11-25T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.482012 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.482053 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.482064 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.482083 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.482094 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:43Z","lastTransitionTime":"2025-11-25T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.503627 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:43 crc kubenswrapper[4802]: E1125 16:48:43.503765 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.583747 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.583788 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.583800 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.583826 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.583838 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:43Z","lastTransitionTime":"2025-11-25T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.685401 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.685456 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.685467 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.685486 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.685500 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:43Z","lastTransitionTime":"2025-11-25T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.788866 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.788921 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.788939 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.788962 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.788981 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:43Z","lastTransitionTime":"2025-11-25T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.891548 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.891629 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.891653 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.891684 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.891707 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:43Z","lastTransitionTime":"2025-11-25T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.994341 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.994417 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.994436 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.994872 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:43 crc kubenswrapper[4802]: I1125 16:48:43.994936 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:43Z","lastTransitionTime":"2025-11-25T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.098369 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.098565 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.098584 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.098647 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.098665 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:44Z","lastTransitionTime":"2025-11-25T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.201403 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.201479 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.201503 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.201534 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.201583 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:44Z","lastTransitionTime":"2025-11-25T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.304795 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.304835 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.304847 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.304865 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.304878 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:44Z","lastTransitionTime":"2025-11-25T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.406811 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.406853 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.406864 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.406879 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.406890 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:44Z","lastTransitionTime":"2025-11-25T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.504262 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.504262 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:44 crc kubenswrapper[4802]: E1125 16:48:44.504403 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.504283 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:44 crc kubenswrapper[4802]: E1125 16:48:44.504481 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:44 crc kubenswrapper[4802]: E1125 16:48:44.504636 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.509361 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.509401 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.509413 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.509427 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.509438 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:44Z","lastTransitionTime":"2025-11-25T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.611714 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.611750 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.611758 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.611774 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.611784 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:44Z","lastTransitionTime":"2025-11-25T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.714500 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.714549 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.714557 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.714578 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.714590 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:44Z","lastTransitionTime":"2025-11-25T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.816927 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.816975 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.816986 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.817003 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.817013 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:44Z","lastTransitionTime":"2025-11-25T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.919582 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.919624 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.919635 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.919650 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.919662 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:44Z","lastTransitionTime":"2025-11-25T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.987401 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.991538 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.991560 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.991579 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 16:48:44 crc kubenswrapper[4802]: I1125 16:48:44.991588 4802 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T16:48:44Z","lastTransitionTime":"2025-11-25T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.028174 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc"] Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.028514 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.030469 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.030670 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.032386 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.032393 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.188209 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/389913e3-3448-44cd-84c6-e2e637882d41-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-spmfc\" (UID: \"389913e3-3448-44cd-84c6-e2e637882d41\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.188269 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/389913e3-3448-44cd-84c6-e2e637882d41-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-spmfc\" (UID: \"389913e3-3448-44cd-84c6-e2e637882d41\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.188311 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/389913e3-3448-44cd-84c6-e2e637882d41-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-spmfc\" (UID: \"389913e3-3448-44cd-84c6-e2e637882d41\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.188341 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/389913e3-3448-44cd-84c6-e2e637882d41-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-spmfc\" (UID: \"389913e3-3448-44cd-84c6-e2e637882d41\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.188360 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/389913e3-3448-44cd-84c6-e2e637882d41-service-ca\") pod \"cluster-version-operator-5c965bbfc6-spmfc\" (UID: \"389913e3-3448-44cd-84c6-e2e637882d41\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.289188 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/389913e3-3448-44cd-84c6-e2e637882d41-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-spmfc\" (UID: \"389913e3-3448-44cd-84c6-e2e637882d41\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.289250 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/389913e3-3448-44cd-84c6-e2e637882d41-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-spmfc\" (UID: \"389913e3-3448-44cd-84c6-e2e637882d41\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.289269 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/389913e3-3448-44cd-84c6-e2e637882d41-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-spmfc\" (UID: \"389913e3-3448-44cd-84c6-e2e637882d41\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.289286 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/389913e3-3448-44cd-84c6-e2e637882d41-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-spmfc\" (UID: \"389913e3-3448-44cd-84c6-e2e637882d41\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.289301 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/389913e3-3448-44cd-84c6-e2e637882d41-service-ca\") pod \"cluster-version-operator-5c965bbfc6-spmfc\" (UID: \"389913e3-3448-44cd-84c6-e2e637882d41\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.289549 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/389913e3-3448-44cd-84c6-e2e637882d41-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-spmfc\" (UID: \"389913e3-3448-44cd-84c6-e2e637882d41\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.289555 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/389913e3-3448-44cd-84c6-e2e637882d41-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-spmfc\" (UID: \"389913e3-3448-44cd-84c6-e2e637882d41\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.290242 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/389913e3-3448-44cd-84c6-e2e637882d41-service-ca\") pod \"cluster-version-operator-5c965bbfc6-spmfc\" (UID: \"389913e3-3448-44cd-84c6-e2e637882d41\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.295660 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/389913e3-3448-44cd-84c6-e2e637882d41-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-spmfc\" (UID: \"389913e3-3448-44cd-84c6-e2e637882d41\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.304263 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/389913e3-3448-44cd-84c6-e2e637882d41-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-spmfc\" (UID: \"389913e3-3448-44cd-84c6-e2e637882d41\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.340861 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.504335 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:45 crc kubenswrapper[4802]: E1125 16:48:45.504595 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.952976 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" event={"ID":"389913e3-3448-44cd-84c6-e2e637882d41","Type":"ContainerStarted","Data":"2a55f2e8ee4f41ab48bd8e7470c1cac8994827b81247cb0bb45a48ce26d9552c"} Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.953037 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" event={"ID":"389913e3-3448-44cd-84c6-e2e637882d41","Type":"ContainerStarted","Data":"4ba0287c2da612572a982eb98aecd16842186e2f758b7cbecc2b509514c75dfd"} Nov 25 16:48:45 crc kubenswrapper[4802]: I1125 16:48:45.969722 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-spmfc" podStartSLOduration=88.969657303 podStartE2EDuration="1m28.969657303s" podCreationTimestamp="2025-11-25 16:47:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:48:45.968584375 +0000 UTC m=+109.112931621" watchObservedRunningTime="2025-11-25 16:48:45.969657303 +0000 UTC m=+109.114004489" Nov 25 16:48:46 crc kubenswrapper[4802]: I1125 16:48:46.503506 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:46 crc kubenswrapper[4802]: I1125 16:48:46.503568 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:46 crc kubenswrapper[4802]: E1125 16:48:46.503627 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:46 crc kubenswrapper[4802]: E1125 16:48:46.503704 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:46 crc kubenswrapper[4802]: I1125 16:48:46.503568 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:46 crc kubenswrapper[4802]: E1125 16:48:46.503847 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:47 crc kubenswrapper[4802]: I1125 16:48:47.503651 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:47 crc kubenswrapper[4802]: E1125 16:48:47.504560 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:48 crc kubenswrapper[4802]: I1125 16:48:48.504390 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:48 crc kubenswrapper[4802]: E1125 16:48:48.504814 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:48 crc kubenswrapper[4802]: I1125 16:48:48.504469 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:48 crc kubenswrapper[4802]: E1125 16:48:48.504894 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:48 crc kubenswrapper[4802]: I1125 16:48:48.504438 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:48 crc kubenswrapper[4802]: E1125 16:48:48.504961 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:49 crc kubenswrapper[4802]: I1125 16:48:49.504076 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:49 crc kubenswrapper[4802]: E1125 16:48:49.504243 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:50 crc kubenswrapper[4802]: I1125 16:48:50.504362 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:50 crc kubenswrapper[4802]: I1125 16:48:50.504389 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:50 crc kubenswrapper[4802]: E1125 16:48:50.504487 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:50 crc kubenswrapper[4802]: I1125 16:48:50.504535 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:50 crc kubenswrapper[4802]: E1125 16:48:50.504667 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:50 crc kubenswrapper[4802]: E1125 16:48:50.504783 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:51 crc kubenswrapper[4802]: I1125 16:48:51.504604 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:51 crc kubenswrapper[4802]: E1125 16:48:51.504739 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:52 crc kubenswrapper[4802]: I1125 16:48:52.503561 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:52 crc kubenswrapper[4802]: I1125 16:48:52.503593 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:52 crc kubenswrapper[4802]: I1125 16:48:52.503581 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:52 crc kubenswrapper[4802]: E1125 16:48:52.503715 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:52 crc kubenswrapper[4802]: E1125 16:48:52.503827 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:52 crc kubenswrapper[4802]: E1125 16:48:52.503902 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:52 crc kubenswrapper[4802]: I1125 16:48:52.504653 4802 scope.go:117] "RemoveContainer" containerID="88787834da7629acd6521f977ee73dec4231d0b4de6a3bb8f4dad83d203b4221" Nov 25 16:48:52 crc kubenswrapper[4802]: I1125 16:48:52.975481 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovnkube-controller/3.log" Nov 25 16:48:52 crc kubenswrapper[4802]: I1125 16:48:52.978082 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerStarted","Data":"7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6"} Nov 25 16:48:52 crc kubenswrapper[4802]: I1125 16:48:52.978568 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:48:52 crc kubenswrapper[4802]: I1125 16:48:52.979366 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hmbvn_97e822d6-58fe-41f4-b08b-3c9b42273307/kube-multus/1.log" Nov 25 16:48:52 crc kubenswrapper[4802]: I1125 16:48:52.979825 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hmbvn_97e822d6-58fe-41f4-b08b-3c9b42273307/kube-multus/0.log" Nov 25 16:48:52 crc kubenswrapper[4802]: I1125 16:48:52.979873 4802 generic.go:334] "Generic (PLEG): container finished" podID="97e822d6-58fe-41f4-b08b-3c9b42273307" containerID="e0e948183d9361bf5f6831d25601431a1de0662987928a7277917451064c0295" exitCode=1 Nov 25 16:48:52 crc kubenswrapper[4802]: I1125 16:48:52.979900 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hmbvn" event={"ID":"97e822d6-58fe-41f4-b08b-3c9b42273307","Type":"ContainerDied","Data":"e0e948183d9361bf5f6831d25601431a1de0662987928a7277917451064c0295"} Nov 25 16:48:52 crc kubenswrapper[4802]: I1125 16:48:52.979922 4802 scope.go:117] "RemoveContainer" containerID="ce2d7addfca2ba375cd1f9a2c618d3e279a2a2f7739d2ef53083a910134cbd8b" Nov 25 16:48:52 crc kubenswrapper[4802]: I1125 16:48:52.980200 4802 scope.go:117] "RemoveContainer" containerID="e0e948183d9361bf5f6831d25601431a1de0662987928a7277917451064c0295" Nov 25 16:48:52 crc kubenswrapper[4802]: E1125 16:48:52.980314 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-hmbvn_openshift-multus(97e822d6-58fe-41f4-b08b-3c9b42273307)\"" pod="openshift-multus/multus-hmbvn" podUID="97e822d6-58fe-41f4-b08b-3c9b42273307" Nov 25 16:48:53 crc kubenswrapper[4802]: I1125 16:48:53.012340 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podStartSLOduration=95.012323196 podStartE2EDuration="1m35.012323196s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:48:53.012154381 +0000 UTC m=+116.156501567" watchObservedRunningTime="2025-11-25 16:48:53.012323196 +0000 UTC m=+116.156670372" Nov 25 16:48:53 crc kubenswrapper[4802]: I1125 16:48:53.217235 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-6ft2z"] Nov 25 16:48:53 crc kubenswrapper[4802]: I1125 16:48:53.217368 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:53 crc kubenswrapper[4802]: E1125 16:48:53.217473 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:53 crc kubenswrapper[4802]: I1125 16:48:53.504412 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:53 crc kubenswrapper[4802]: E1125 16:48:53.504599 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:53 crc kubenswrapper[4802]: I1125 16:48:53.984086 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hmbvn_97e822d6-58fe-41f4-b08b-3c9b42273307/kube-multus/1.log" Nov 25 16:48:54 crc kubenswrapper[4802]: I1125 16:48:54.504293 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:54 crc kubenswrapper[4802]: I1125 16:48:54.504350 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:54 crc kubenswrapper[4802]: I1125 16:48:54.504343 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:54 crc kubenswrapper[4802]: E1125 16:48:54.504418 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:54 crc kubenswrapper[4802]: E1125 16:48:54.504581 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:54 crc kubenswrapper[4802]: E1125 16:48:54.504625 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:55 crc kubenswrapper[4802]: I1125 16:48:55.503774 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:55 crc kubenswrapper[4802]: E1125 16:48:55.503947 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:56 crc kubenswrapper[4802]: I1125 16:48:56.503686 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:56 crc kubenswrapper[4802]: I1125 16:48:56.503777 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:56 crc kubenswrapper[4802]: E1125 16:48:56.503855 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:56 crc kubenswrapper[4802]: I1125 16:48:56.503898 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:56 crc kubenswrapper[4802]: E1125 16:48:56.504014 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:56 crc kubenswrapper[4802]: E1125 16:48:56.504047 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:57 crc kubenswrapper[4802]: I1125 16:48:57.503875 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:57 crc kubenswrapper[4802]: E1125 16:48:57.505494 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:57 crc kubenswrapper[4802]: E1125 16:48:57.555735 4802 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 25 16:48:57 crc kubenswrapper[4802]: E1125 16:48:57.598060 4802 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 16:48:58 crc kubenswrapper[4802]: I1125 16:48:58.503568 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:48:58 crc kubenswrapper[4802]: I1125 16:48:58.503658 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:48:58 crc kubenswrapper[4802]: E1125 16:48:58.503810 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:48:58 crc kubenswrapper[4802]: I1125 16:48:58.503827 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:48:58 crc kubenswrapper[4802]: E1125 16:48:58.504040 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:48:58 crc kubenswrapper[4802]: E1125 16:48:58.504242 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:48:59 crc kubenswrapper[4802]: I1125 16:48:59.504023 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:48:59 crc kubenswrapper[4802]: E1125 16:48:59.504426 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:48:59 crc kubenswrapper[4802]: I1125 16:48:59.729775 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:49:00 crc kubenswrapper[4802]: I1125 16:49:00.503907 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:49:00 crc kubenswrapper[4802]: E1125 16:49:00.504288 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:49:00 crc kubenswrapper[4802]: I1125 16:49:00.504017 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:49:00 crc kubenswrapper[4802]: E1125 16:49:00.504789 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:49:00 crc kubenswrapper[4802]: I1125 16:49:00.504017 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:49:00 crc kubenswrapper[4802]: E1125 16:49:00.505024 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:49:01 crc kubenswrapper[4802]: I1125 16:49:01.503885 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:49:01 crc kubenswrapper[4802]: E1125 16:49:01.504031 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:49:02 crc kubenswrapper[4802]: I1125 16:49:02.504536 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:49:02 crc kubenswrapper[4802]: I1125 16:49:02.504559 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:49:02 crc kubenswrapper[4802]: E1125 16:49:02.504724 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:49:02 crc kubenswrapper[4802]: I1125 16:49:02.504837 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:49:02 crc kubenswrapper[4802]: E1125 16:49:02.504960 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:49:02 crc kubenswrapper[4802]: E1125 16:49:02.505154 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:49:02 crc kubenswrapper[4802]: E1125 16:49:02.599876 4802 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 16:49:03 crc kubenswrapper[4802]: I1125 16:49:03.503987 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:49:03 crc kubenswrapper[4802]: E1125 16:49:03.504162 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:49:04 crc kubenswrapper[4802]: I1125 16:49:04.504164 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:49:04 crc kubenswrapper[4802]: I1125 16:49:04.504243 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:49:04 crc kubenswrapper[4802]: E1125 16:49:04.504323 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:49:04 crc kubenswrapper[4802]: I1125 16:49:04.504261 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:49:04 crc kubenswrapper[4802]: E1125 16:49:04.504397 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:49:04 crc kubenswrapper[4802]: E1125 16:49:04.504487 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:49:05 crc kubenswrapper[4802]: I1125 16:49:05.504114 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:49:05 crc kubenswrapper[4802]: E1125 16:49:05.504411 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:49:06 crc kubenswrapper[4802]: I1125 16:49:06.504323 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:49:06 crc kubenswrapper[4802]: I1125 16:49:06.504392 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:49:06 crc kubenswrapper[4802]: I1125 16:49:06.504341 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:49:06 crc kubenswrapper[4802]: E1125 16:49:06.504581 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:49:06 crc kubenswrapper[4802]: E1125 16:49:06.504685 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:49:06 crc kubenswrapper[4802]: E1125 16:49:06.504772 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:49:06 crc kubenswrapper[4802]: I1125 16:49:06.505294 4802 scope.go:117] "RemoveContainer" containerID="e0e948183d9361bf5f6831d25601431a1de0662987928a7277917451064c0295" Nov 25 16:49:07 crc kubenswrapper[4802]: I1125 16:49:07.027882 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hmbvn_97e822d6-58fe-41f4-b08b-3c9b42273307/kube-multus/1.log" Nov 25 16:49:07 crc kubenswrapper[4802]: I1125 16:49:07.028309 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hmbvn" event={"ID":"97e822d6-58fe-41f4-b08b-3c9b42273307","Type":"ContainerStarted","Data":"ade8eae2c0adbaed7320d8fd691aabc2c6cf3d3c6b38153dedfccd37ae7a67f8"} Nov 25 16:49:07 crc kubenswrapper[4802]: I1125 16:49:07.504550 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:49:07 crc kubenswrapper[4802]: E1125 16:49:07.505548 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:49:07 crc kubenswrapper[4802]: E1125 16:49:07.600614 4802 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 16:49:08 crc kubenswrapper[4802]: I1125 16:49:08.504282 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:49:08 crc kubenswrapper[4802]: I1125 16:49:08.504353 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:49:08 crc kubenswrapper[4802]: I1125 16:49:08.504460 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:49:08 crc kubenswrapper[4802]: E1125 16:49:08.504551 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:49:08 crc kubenswrapper[4802]: E1125 16:49:08.504682 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:49:08 crc kubenswrapper[4802]: E1125 16:49:08.504807 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:49:09 crc kubenswrapper[4802]: I1125 16:49:09.504256 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:49:09 crc kubenswrapper[4802]: E1125 16:49:09.504462 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:49:10 crc kubenswrapper[4802]: I1125 16:49:10.503923 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:49:10 crc kubenswrapper[4802]: I1125 16:49:10.503966 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:49:10 crc kubenswrapper[4802]: I1125 16:49:10.504003 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:49:10 crc kubenswrapper[4802]: E1125 16:49:10.504066 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:49:10 crc kubenswrapper[4802]: E1125 16:49:10.504241 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:49:10 crc kubenswrapper[4802]: E1125 16:49:10.504336 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:49:11 crc kubenswrapper[4802]: I1125 16:49:11.504159 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:49:11 crc kubenswrapper[4802]: E1125 16:49:11.504443 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 16:49:12 crc kubenswrapper[4802]: I1125 16:49:12.504210 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:49:12 crc kubenswrapper[4802]: I1125 16:49:12.504305 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:49:12 crc kubenswrapper[4802]: I1125 16:49:12.504330 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:49:12 crc kubenswrapper[4802]: E1125 16:49:12.504613 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6ft2z" podUID="cd9aa596-76c2-468c-b732-d85bc1f16070" Nov 25 16:49:12 crc kubenswrapper[4802]: E1125 16:49:12.504751 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 16:49:12 crc kubenswrapper[4802]: E1125 16:49:12.504938 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 16:49:13 crc kubenswrapper[4802]: I1125 16:49:13.504425 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:49:13 crc kubenswrapper[4802]: I1125 16:49:13.506676 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 16:49:13 crc kubenswrapper[4802]: I1125 16:49:13.506924 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 16:49:14 crc kubenswrapper[4802]: I1125 16:49:14.504510 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:49:14 crc kubenswrapper[4802]: I1125 16:49:14.504550 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:49:14 crc kubenswrapper[4802]: I1125 16:49:14.504791 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:49:14 crc kubenswrapper[4802]: I1125 16:49:14.508684 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 16:49:14 crc kubenswrapper[4802]: I1125 16:49:14.509616 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 16:49:14 crc kubenswrapper[4802]: I1125 16:49:14.509837 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 16:49:14 crc kubenswrapper[4802]: I1125 16:49:14.509898 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.398853 4802 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.432446 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.432990 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.435987 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.435989 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.436724 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.436989 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.437058 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.437112 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.437151 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.437409 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.437711 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.437978 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.438460 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hzx2q"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.439045 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.439572 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j2t9z"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.440337 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.440665 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-j2t9z" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.440723 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bv6b6"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.441362 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.442463 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mpcc"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.443108 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mpcc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.444188 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.444399 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.450093 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sccr2"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.450631 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.456071 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-p7zvz"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.457055 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.459470 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2prtt"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.460040 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.461569 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9qz4w"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.462396 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9qz4w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.462684 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-smpjk"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.463616 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.464963 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.465299 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.465554 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.465691 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.474632 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.475031 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.477732 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.487142 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.487648 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.488634 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.488887 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.489063 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.490332 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.490491 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.490600 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.491753 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-nr4zg"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.492806 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.492859 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.493207 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.493248 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.493844 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.494006 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516280 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516347 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/54c5e683-04bc-4241-8d13-17b415237707-encryption-config\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516373 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c673008d-695e-44a5-8b6d-1d7c585b5b72-machine-approver-tls\") pod \"machine-approver-56656f9798-2wcfc\" (UID: \"c673008d-695e-44a5-8b6d-1d7c585b5b72\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516397 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-node-pullsecrets\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516423 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-image-import-ca\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516465 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v57bl\" (UniqueName: \"kubernetes.io/projected/ea30ecf0-8806-4c5a-9ff6-2738ca72e50b-kube-api-access-v57bl\") pod \"machine-api-operator-5694c8668f-j2t9z\" (UID: \"ea30ecf0-8806-4c5a-9ff6-2738ca72e50b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j2t9z" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516494 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwl2t\" (UniqueName: \"kubernetes.io/projected/54c5e683-04bc-4241-8d13-17b415237707-kube-api-access-zwl2t\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516522 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/54c5e683-04bc-4241-8d13-17b415237707-etcd-client\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516543 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54c5e683-04bc-4241-8d13-17b415237707-serving-cert\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516566 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-client-ca\") pod \"route-controller-manager-6576b87f9c-zvb9w\" (UID: \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516591 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-config\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516613 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z46ch\" (UniqueName: \"kubernetes.io/projected/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-kube-api-access-z46ch\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516640 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54c5e683-04bc-4241-8d13-17b415237707-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516662 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49937863-524b-4ee0-b3c4-6522995947c4-config\") pod \"controller-manager-879f6c89f-bv6b6\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516690 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-serving-cert\") pod \"route-controller-manager-6576b87f9c-zvb9w\" (UID: \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516720 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/49937863-524b-4ee0-b3c4-6522995947c4-client-ca\") pod \"controller-manager-879f6c89f-bv6b6\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516745 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c673008d-695e-44a5-8b6d-1d7c585b5b72-auth-proxy-config\") pod \"machine-approver-56656f9798-2wcfc\" (UID: \"c673008d-695e-44a5-8b6d-1d7c585b5b72\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516764 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ea30ecf0-8806-4c5a-9ff6-2738ca72e50b-images\") pod \"machine-api-operator-5694c8668f-j2t9z\" (UID: \"ea30ecf0-8806-4c5a-9ff6-2738ca72e50b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j2t9z" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516811 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/54c5e683-04bc-4241-8d13-17b415237707-audit-dir\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516841 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-etcd-serving-ca\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516868 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/49937863-524b-4ee0-b3c4-6522995947c4-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-bv6b6\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517469 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-etcd-client\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517501 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea30ecf0-8806-4c5a-9ff6-2738ca72e50b-config\") pod \"machine-api-operator-5694c8668f-j2t9z\" (UID: \"ea30ecf0-8806-4c5a-9ff6-2738ca72e50b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j2t9z" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517527 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ea30ecf0-8806-4c5a-9ff6-2738ca72e50b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j2t9z\" (UID: \"ea30ecf0-8806-4c5a-9ff6-2738ca72e50b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j2t9z" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517569 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-audit-dir\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517597 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n8jz\" (UniqueName: \"kubernetes.io/projected/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-kube-api-access-4n8jz\") pod \"route-controller-manager-6576b87f9c-zvb9w\" (UID: \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517624 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c673008d-695e-44a5-8b6d-1d7c585b5b72-config\") pod \"machine-approver-56656f9798-2wcfc\" (UID: \"c673008d-695e-44a5-8b6d-1d7c585b5b72\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517648 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-serving-cert\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517699 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/54c5e683-04bc-4241-8d13-17b415237707-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517729 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-encryption-config\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517764 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-config\") pod \"route-controller-manager-6576b87f9c-zvb9w\" (UID: \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517799 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49937863-524b-4ee0-b3c4-6522995947c4-serving-cert\") pod \"controller-manager-879f6c89f-bv6b6\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517828 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4x99\" (UniqueName: \"kubernetes.io/projected/49937863-524b-4ee0-b3c4-6522995947c4-kube-api-access-t4x99\") pod \"controller-manager-879f6c89f-bv6b6\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517861 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7ljt\" (UniqueName: \"kubernetes.io/projected/c673008d-695e-44a5-8b6d-1d7c585b5b72-kube-api-access-n7ljt\") pod \"machine-approver-56656f9798-2wcfc\" (UID: \"c673008d-695e-44a5-8b6d-1d7c585b5b72\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517906 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/54c5e683-04bc-4241-8d13-17b415237707-audit-policies\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517938 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-audit\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516862 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.518146 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.518177 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.516998 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.518352 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.518375 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517014 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517073 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517149 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517587 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.517630 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.530545 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.530780 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.530820 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.530926 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.531139 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.531179 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.531265 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.531388 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.531402 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.531471 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.531547 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.531670 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.531683 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.531695 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.531761 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.531952 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.532066 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.533197 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.533296 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.533377 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.533460 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.533565 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.533643 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.533751 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.533780 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.533948 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.533972 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.534193 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.534302 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.534343 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.534384 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.534444 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.534529 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.534609 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.534664 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.534740 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.534829 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.534919 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.535059 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.539114 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.539244 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.539312 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.539383 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.539581 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.541651 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.544468 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-79vhf"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.544973 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-79vhf" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.547205 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.547547 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.550200 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.550824 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9tt57"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.551073 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kst7h"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.551093 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.551462 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.551591 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.551681 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.551879 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9tt57" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.551954 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.552324 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.552418 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.552428 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.552437 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.552520 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.552566 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.552614 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.554986 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.557785 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.559999 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-nh2kl"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.560614 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.561034 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zpl2d"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.561355 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zpl2d" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.561426 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jzd75"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.561865 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jzd75" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.564428 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmm9"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.564768 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sl8lb"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.564789 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.565102 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sl8lb" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.565196 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmm9" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.565232 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.565313 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-hk4bx"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.565436 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.569109 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-hk4bx" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.572901 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.573418 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-852rn"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.575344 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9lzl9"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.575966 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.576970 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9lzl9" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.577888 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-852rn" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.590405 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-kk8nf"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.590711 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.591403 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.591851 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-nmzxv"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.592307 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-nmzxv" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.592610 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kk8nf" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.592809 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.593056 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mpcc"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.593233 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.594668 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.595311 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.597328 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.597582 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zjrnl"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.599309 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-z8dgw"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.599658 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-z8dgw" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.599836 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zjrnl" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.601292 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6hwbq"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.601339 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.601968 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6hwbq" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.602276 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ppfnt"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.602930 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-ppfnt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.603190 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9qz4w"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.607519 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.608077 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.608493 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.608988 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.610187 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.610885 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.610958 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cr9nk"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.611370 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.613390 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j2t9z"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.614861 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2prtt"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.615681 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.626390 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-nh2kl"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.626444 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-p7zvz"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.626456 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-v6gj4"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.627202 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-v6gj4" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.630065 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jzd75"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.630106 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-cfw4c"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.632160 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ea30ecf0-8806-4c5a-9ff6-2738ca72e50b-images\") pod \"machine-api-operator-5694c8668f-j2t9z\" (UID: \"ea30ecf0-8806-4c5a-9ff6-2738ca72e50b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j2t9z" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.632302 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w9gv\" (UniqueName: \"kubernetes.io/projected/7a17670c-87d6-4fc7-b197-6817f4467377-kube-api-access-8w9gv\") pod \"cluster-image-registry-operator-dc59b4c8b-2jkt6\" (UID: \"7a17670c-87d6-4fc7-b197-6817f4467377\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.632956 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/54c5e683-04bc-4241-8d13-17b415237707-audit-dir\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.633050 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-etcd-serving-ca\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.633143 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sb4p\" (UniqueName: \"kubernetes.io/projected/df33868c-7e7c-4d67-b478-40ee1d7ef69d-kube-api-access-2sb4p\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.633228 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-audit-policies\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.633304 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/49937863-524b-4ee0-b3c4-6522995947c4-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-bv6b6\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.633373 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq54s\" (UniqueName: \"kubernetes.io/projected/9af7072b-a939-4cdb-813f-1dccb589763a-kube-api-access-sq54s\") pod \"cluster-samples-operator-665b6dd947-9qz4w\" (UID: \"9af7072b-a939-4cdb-813f-1dccb589763a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9qz4w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.633449 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f7d5d65a-b6dd-45dd-9f35-5e02a915a64f-stats-auth\") pod \"router-default-5444994796-nr4zg\" (UID: \"f7d5d65a-b6dd-45dd-9f35-5e02a915a64f\") " pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.633549 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.633650 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-etcd-client\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.633744 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea30ecf0-8806-4c5a-9ff6-2738ca72e50b-config\") pod \"machine-api-operator-5694c8668f-j2t9z\" (UID: \"ea30ecf0-8806-4c5a-9ff6-2738ca72e50b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j2t9z" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.633820 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ea30ecf0-8806-4c5a-9ff6-2738ca72e50b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j2t9z\" (UID: \"ea30ecf0-8806-4c5a-9ff6-2738ca72e50b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j2t9z" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.633906 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/df33868c-7e7c-4d67-b478-40ee1d7ef69d-console-serving-cert\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.634003 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91ee6068-d199-4b76-81fa-897155d5a38d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7mpcc\" (UID: \"91ee6068-d199-4b76-81fa-897155d5a38d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mpcc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.634085 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62710ae2-1843-4880-8444-b501a206e145-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2prtt\" (UID: \"62710ae2-1843-4880-8444-b501a206e145\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.634191 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f7d5d65a-b6dd-45dd-9f35-5e02a915a64f-metrics-certs\") pod \"router-default-5444994796-nr4zg\" (UID: \"f7d5d65a-b6dd-45dd-9f35-5e02a915a64f\") " pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.634288 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.634431 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-audit-dir\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.634512 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ea30ecf0-8806-4c5a-9ff6-2738ca72e50b-images\") pod \"machine-api-operator-5694c8668f-j2t9z\" (UID: \"ea30ecf0-8806-4c5a-9ff6-2738ca72e50b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j2t9z" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.634310 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-cfw4c" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.634533 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n8jz\" (UniqueName: \"kubernetes.io/projected/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-kube-api-access-4n8jz\") pod \"route-controller-manager-6576b87f9c-zvb9w\" (UID: \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.634702 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c673008d-695e-44a5-8b6d-1d7c585b5b72-config\") pod \"machine-approver-56656f9798-2wcfc\" (UID: \"c673008d-695e-44a5-8b6d-1d7c585b5b72\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.634772 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7a17670c-87d6-4fc7-b197-6817f4467377-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2jkt6\" (UID: \"7a17670c-87d6-4fc7-b197-6817f4467377\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.633593 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/54c5e683-04bc-4241-8d13-17b415237707-audit-dir\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.634912 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/54c5e683-04bc-4241-8d13-17b415237707-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.634000 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.635019 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sl8lb"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.635039 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.635105 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-serving-cert\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.635399 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f7d5d65a-b6dd-45dd-9f35-5e02a915a64f-default-certificate\") pod \"router-default-5444994796-nr4zg\" (UID: \"f7d5d65a-b6dd-45dd-9f35-5e02a915a64f\") " pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.635506 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-encryption-config\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.635589 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-config\") pod \"route-controller-manager-6576b87f9c-zvb9w\" (UID: \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.635647 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/49937863-524b-4ee0-b3c4-6522995947c4-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-bv6b6\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.634443 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-etcd-serving-ca\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.635197 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea30ecf0-8806-4c5a-9ff6-2738ca72e50b-config\") pod \"machine-api-operator-5694c8668f-j2t9z\" (UID: \"ea30ecf0-8806-4c5a-9ff6-2738ca72e50b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j2t9z" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.635260 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-audit-dir\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.635791 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/54c5e683-04bc-4241-8d13-17b415237707-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.635800 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49937863-524b-4ee0-b3c4-6522995947c4-serving-cert\") pod \"controller-manager-879f6c89f-bv6b6\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.635861 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4x99\" (UniqueName: \"kubernetes.io/projected/49937863-524b-4ee0-b3c4-6522995947c4-kube-api-access-t4x99\") pod \"controller-manager-879f6c89f-bv6b6\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.635883 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7ljt\" (UniqueName: \"kubernetes.io/projected/c673008d-695e-44a5-8b6d-1d7c585b5b72-kube-api-access-n7ljt\") pod \"machine-approver-56656f9798-2wcfc\" (UID: \"c673008d-695e-44a5-8b6d-1d7c585b5b72\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.635920 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62710ae2-1843-4880-8444-b501a206e145-serving-cert\") pod \"authentication-operator-69f744f599-2prtt\" (UID: \"62710ae2-1843-4880-8444-b501a206e145\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.635921 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c673008d-695e-44a5-8b6d-1d7c585b5b72-config\") pod \"machine-approver-56656f9798-2wcfc\" (UID: \"c673008d-695e-44a5-8b6d-1d7c585b5b72\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.635940 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636020 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/54c5e683-04bc-4241-8d13-17b415237707-audit-policies\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636060 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7d5d65a-b6dd-45dd-9f35-5e02a915a64f-service-ca-bundle\") pod \"router-default-5444994796-nr4zg\" (UID: \"f7d5d65a-b6dd-45dd-9f35-5e02a915a64f\") " pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636083 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/df33868c-7e7c-4d67-b478-40ee1d7ef69d-oauth-serving-cert\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636105 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636211 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-audit\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636248 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0-available-featuregates\") pod \"openshift-config-operator-7777fb866f-smpjk\" (UID: \"b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636278 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636315 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/7a17670c-87d6-4fc7-b197-6817f4467377-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2jkt6\" (UID: \"7a17670c-87d6-4fc7-b197-6817f4467377\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636341 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqwm7\" (UniqueName: \"kubernetes.io/projected/62710ae2-1843-4880-8444-b501a206e145-kube-api-access-fqwm7\") pod \"authentication-operator-69f744f599-2prtt\" (UID: \"62710ae2-1843-4880-8444-b501a206e145\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636365 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636389 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfvrl\" (UniqueName: \"kubernetes.io/projected/b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0-kube-api-access-qfvrl\") pod \"openshift-config-operator-7777fb866f-smpjk\" (UID: \"b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636417 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636444 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/54c5e683-04bc-4241-8d13-17b415237707-encryption-config\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636470 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k25r\" (UniqueName: \"kubernetes.io/projected/91ee6068-d199-4b76-81fa-897155d5a38d-kube-api-access-6k25r\") pod \"openshift-apiserver-operator-796bbdcf4f-7mpcc\" (UID: \"91ee6068-d199-4b76-81fa-897155d5a38d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mpcc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636495 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62710ae2-1843-4880-8444-b501a206e145-config\") pod \"authentication-operator-69f744f599-2prtt\" (UID: \"62710ae2-1843-4880-8444-b501a206e145\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636522 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-node-pullsecrets\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636547 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c673008d-695e-44a5-8b6d-1d7c585b5b72-machine-approver-tls\") pod \"machine-approver-56656f9798-2wcfc\" (UID: \"c673008d-695e-44a5-8b6d-1d7c585b5b72\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636566 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/54c5e683-04bc-4241-8d13-17b415237707-audit-policies\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636579 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-image-import-ca\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636607 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/63972ba9-04aa-42e8-a91c-f27796ab40f4-audit-dir\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636631 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636653 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0-serving-cert\") pod \"openshift-config-operator-7777fb866f-smpjk\" (UID: \"b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636694 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v57bl\" (UniqueName: \"kubernetes.io/projected/ea30ecf0-8806-4c5a-9ff6-2738ca72e50b-kube-api-access-v57bl\") pod \"machine-api-operator-5694c8668f-j2t9z\" (UID: \"ea30ecf0-8806-4c5a-9ff6-2738ca72e50b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j2t9z" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636717 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91ee6068-d199-4b76-81fa-897155d5a38d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7mpcc\" (UID: \"91ee6068-d199-4b76-81fa-897155d5a38d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mpcc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636741 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm5mh\" (UniqueName: \"kubernetes.io/projected/f7d5d65a-b6dd-45dd-9f35-5e02a915a64f-kube-api-access-lm5mh\") pod \"router-default-5444994796-nr4zg\" (UID: \"f7d5d65a-b6dd-45dd-9f35-5e02a915a64f\") " pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636772 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwl2t\" (UniqueName: \"kubernetes.io/projected/54c5e683-04bc-4241-8d13-17b415237707-kube-api-access-zwl2t\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636801 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df33868c-7e7c-4d67-b478-40ee1d7ef69d-trusted-ca-bundle\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636826 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54c5e683-04bc-4241-8d13-17b415237707-serving-cert\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636849 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/54c5e683-04bc-4241-8d13-17b415237707-etcd-client\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636872 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-client-ca\") pod \"route-controller-manager-6576b87f9c-zvb9w\" (UID: \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636896 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.636979 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-audit\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.637063 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-config\") pod \"route-controller-manager-6576b87f9c-zvb9w\" (UID: \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.637416 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-config\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.637487 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z46ch\" (UniqueName: \"kubernetes.io/projected/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-kube-api-access-z46ch\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.637519 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9af7072b-a939-4cdb-813f-1dccb589763a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9qz4w\" (UID: \"9af7072b-a939-4cdb-813f-1dccb589763a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9qz4w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.637544 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.637876 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.638085 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-client-ca\") pod \"route-controller-manager-6576b87f9c-zvb9w\" (UID: \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.638989 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-image-import-ca\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.638978 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-node-pullsecrets\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.639253 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/df33868c-7e7c-4d67-b478-40ee1d7ef69d-console-oauth-config\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.639294 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54c5e683-04bc-4241-8d13-17b415237707-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.639685 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54c5e683-04bc-4241-8d13-17b415237707-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.639730 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49937863-524b-4ee0-b3c4-6522995947c4-config\") pod \"controller-manager-879f6c89f-bv6b6\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.639765 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7a17670c-87d6-4fc7-b197-6817f4467377-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2jkt6\" (UID: \"7a17670c-87d6-4fc7-b197-6817f4467377\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.639784 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/df33868c-7e7c-4d67-b478-40ee1d7ef69d-console-config\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.639808 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.641952 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62710ae2-1843-4880-8444-b501a206e145-service-ca-bundle\") pod \"authentication-operator-69f744f599-2prtt\" (UID: \"62710ae2-1843-4880-8444-b501a206e145\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.642026 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-serving-cert\") pod \"route-controller-manager-6576b87f9c-zvb9w\" (UID: \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.642073 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8npcc\" (UniqueName: \"kubernetes.io/projected/63972ba9-04aa-42e8-a91c-f27796ab40f4-kube-api-access-8npcc\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.643340 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/49937863-524b-4ee0-b3c4-6522995947c4-client-ca\") pod \"controller-manager-879f6c89f-bv6b6\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.643384 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/df33868c-7e7c-4d67-b478-40ee1d7ef69d-service-ca\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.643405 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.643444 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c673008d-695e-44a5-8b6d-1d7c585b5b72-auth-proxy-config\") pod \"machine-approver-56656f9798-2wcfc\" (UID: \"c673008d-695e-44a5-8b6d-1d7c585b5b72\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.644027 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/49937863-524b-4ee0-b3c4-6522995947c4-client-ca\") pod \"controller-manager-879f6c89f-bv6b6\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.644190 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bv6b6"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.644216 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c673008d-695e-44a5-8b6d-1d7c585b5b72-auth-proxy-config\") pod \"machine-approver-56656f9798-2wcfc\" (UID: \"c673008d-695e-44a5-8b6d-1d7c585b5b72\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.646858 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49937863-524b-4ee0-b3c4-6522995947c4-serving-cert\") pod \"controller-manager-879f6c89f-bv6b6\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.646873 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c673008d-695e-44a5-8b6d-1d7c585b5b72-machine-approver-tls\") pod \"machine-approver-56656f9798-2wcfc\" (UID: \"c673008d-695e-44a5-8b6d-1d7c585b5b72\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.649685 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49937863-524b-4ee0-b3c4-6522995947c4-config\") pod \"controller-manager-879f6c89f-bv6b6\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.651445 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.652177 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-serving-cert\") pod \"route-controller-manager-6576b87f9c-zvb9w\" (UID: \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.652989 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-config\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.655616 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-smpjk"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.657519 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-852rn"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.660759 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.667578 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hzx2q"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.668755 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-etcd-client\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.668925 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ea30ecf0-8806-4c5a-9ff6-2738ca72e50b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j2t9z\" (UID: \"ea30ecf0-8806-4c5a-9ff6-2738ca72e50b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j2t9z" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.669026 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-encryption-config\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.669924 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/54c5e683-04bc-4241-8d13-17b415237707-encryption-config\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.672392 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.672521 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54c5e683-04bc-4241-8d13-17b415237707-serving-cert\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.675172 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.675243 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9lzl9"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.675709 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-serving-cert\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.682111 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/54c5e683-04bc-4241-8d13-17b415237707-etcd-client\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.682244 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-79vhf"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.688238 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9tt57"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.688627 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kst7h"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.692555 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.693947 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmm9"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.695919 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-z8dgw"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.696052 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.697494 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.698623 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zpl2d"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.699649 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-hk4bx"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.700673 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.701671 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6hwbq"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.702678 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sccr2"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.703660 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ppfnt"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.704660 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-s684g"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.706323 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-4sl2t"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.706489 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.707239 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-cfw4c"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.707356 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-4sl2t" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.708155 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-nmzxv"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.709242 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zjrnl"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.710358 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-v6gj4"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.711397 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-kk8nf"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.712609 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cr9nk"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.713502 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.714564 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.715574 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-s684g"] Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.721430 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.735240 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.744027 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62710ae2-1843-4880-8444-b501a206e145-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2prtt\" (UID: \"62710ae2-1843-4880-8444-b501a206e145\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.744072 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f7d5d65a-b6dd-45dd-9f35-5e02a915a64f-metrics-certs\") pod \"router-default-5444994796-nr4zg\" (UID: \"f7d5d65a-b6dd-45dd-9f35-5e02a915a64f\") " pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.744111 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.744905 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7a17670c-87d6-4fc7-b197-6817f4467377-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2jkt6\" (UID: \"7a17670c-87d6-4fc7-b197-6817f4467377\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.745105 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.745395 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62710ae2-1843-4880-8444-b501a206e145-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2prtt\" (UID: \"62710ae2-1843-4880-8444-b501a206e145\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.745191 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f7d5d65a-b6dd-45dd-9f35-5e02a915a64f-default-certificate\") pod \"router-default-5444994796-nr4zg\" (UID: \"f7d5d65a-b6dd-45dd-9f35-5e02a915a64f\") " pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.745883 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62710ae2-1843-4880-8444-b501a206e145-serving-cert\") pod \"authentication-operator-69f744f599-2prtt\" (UID: \"62710ae2-1843-4880-8444-b501a206e145\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.746038 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.746887 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fbb37aa9-5a46-4ef3-9251-748a657e0933-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zpl2d\" (UID: \"fbb37aa9-5a46-4ef3-9251-748a657e0933\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zpl2d" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.746921 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13d7ba8a-bacc-4294-8e56-c3425f284c1a-config\") pod \"kube-controller-manager-operator-78b949d7b-sl8lb\" (UID: \"13d7ba8a-bacc-4294-8e56-c3425f284c1a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sl8lb" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.746943 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/df33868c-7e7c-4d67-b478-40ee1d7ef69d-oauth-serving-cert\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.746968 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747010 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7d5d65a-b6dd-45dd-9f35-5e02a915a64f-service-ca-bundle\") pod \"router-default-5444994796-nr4zg\" (UID: \"f7d5d65a-b6dd-45dd-9f35-5e02a915a64f\") " pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747057 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0-available-featuregates\") pod \"openshift-config-operator-7777fb866f-smpjk\" (UID: \"b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747142 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747174 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfvrl\" (UniqueName: \"kubernetes.io/projected/b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0-kube-api-access-qfvrl\") pod \"openshift-config-operator-7777fb866f-smpjk\" (UID: \"b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747196 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/7a17670c-87d6-4fc7-b197-6817f4467377-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2jkt6\" (UID: \"7a17670c-87d6-4fc7-b197-6817f4467377\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747219 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqwm7\" (UniqueName: \"kubernetes.io/projected/62710ae2-1843-4880-8444-b501a206e145-kube-api-access-fqwm7\") pod \"authentication-operator-69f744f599-2prtt\" (UID: \"62710ae2-1843-4880-8444-b501a206e145\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747236 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747258 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k25r\" (UniqueName: \"kubernetes.io/projected/91ee6068-d199-4b76-81fa-897155d5a38d-kube-api-access-6k25r\") pod \"openshift-apiserver-operator-796bbdcf4f-7mpcc\" (UID: \"91ee6068-d199-4b76-81fa-897155d5a38d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mpcc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747276 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62710ae2-1843-4880-8444-b501a206e145-config\") pod \"authentication-operator-69f744f599-2prtt\" (UID: \"62710ae2-1843-4880-8444-b501a206e145\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747301 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/17422b05-1782-421b-ab25-4ce61d267fc2-config-volume\") pod \"collect-profiles-29401485-28zdh\" (UID: \"17422b05-1782-421b-ab25-4ce61d267fc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747324 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747342 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0-serving-cert\") pod \"openshift-config-operator-7777fb866f-smpjk\" (UID: \"b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747377 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/63972ba9-04aa-42e8-a91c-f27796ab40f4-audit-dir\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747391 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/17422b05-1782-421b-ab25-4ce61d267fc2-secret-volume\") pod \"collect-profiles-29401485-28zdh\" (UID: \"17422b05-1782-421b-ab25-4ce61d267fc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747416 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91ee6068-d199-4b76-81fa-897155d5a38d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7mpcc\" (UID: \"91ee6068-d199-4b76-81fa-897155d5a38d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mpcc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747439 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm5mh\" (UniqueName: \"kubernetes.io/projected/f7d5d65a-b6dd-45dd-9f35-5e02a915a64f-kube-api-access-lm5mh\") pod \"router-default-5444994796-nr4zg\" (UID: \"f7d5d65a-b6dd-45dd-9f35-5e02a915a64f\") " pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747469 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df33868c-7e7c-4d67-b478-40ee1d7ef69d-trusted-ca-bundle\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747492 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747524 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9af7072b-a939-4cdb-813f-1dccb589763a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9qz4w\" (UID: \"9af7072b-a939-4cdb-813f-1dccb589763a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9qz4w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747541 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747564 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/df33868c-7e7c-4d67-b478-40ee1d7ef69d-console-oauth-config\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747584 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2mcn\" (UniqueName: \"kubernetes.io/projected/8a3ebc4f-0491-4dc1-8c5d-adb851b6ef87-kube-api-access-w2mcn\") pod \"downloads-7954f5f757-z8dgw\" (UID: \"8a3ebc4f-0491-4dc1-8c5d-adb851b6ef87\") " pod="openshift-console/downloads-7954f5f757-z8dgw" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747604 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t96bv\" (UniqueName: \"kubernetes.io/projected/17422b05-1782-421b-ab25-4ce61d267fc2-kube-api-access-t96bv\") pod \"collect-profiles-29401485-28zdh\" (UID: \"17422b05-1782-421b-ab25-4ce61d267fc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747620 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/df33868c-7e7c-4d67-b478-40ee1d7ef69d-console-config\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747635 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747654 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7a17670c-87d6-4fc7-b197-6817f4467377-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2jkt6\" (UID: \"7a17670c-87d6-4fc7-b197-6817f4467377\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747676 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62710ae2-1843-4880-8444-b501a206e145-service-ca-bundle\") pod \"authentication-operator-69f744f599-2prtt\" (UID: \"62710ae2-1843-4880-8444-b501a206e145\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747697 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8npcc\" (UniqueName: \"kubernetes.io/projected/63972ba9-04aa-42e8-a91c-f27796ab40f4-kube-api-access-8npcc\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747717 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/df33868c-7e7c-4d67-b478-40ee1d7ef69d-service-ca\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747732 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747749 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbb37aa9-5a46-4ef3-9251-748a657e0933-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zpl2d\" (UID: \"fbb37aa9-5a46-4ef3-9251-748a657e0933\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zpl2d" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747767 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13d7ba8a-bacc-4294-8e56-c3425f284c1a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sl8lb\" (UID: \"13d7ba8a-bacc-4294-8e56-c3425f284c1a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sl8lb" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747847 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w9gv\" (UniqueName: \"kubernetes.io/projected/7a17670c-87d6-4fc7-b197-6817f4467377-kube-api-access-8w9gv\") pod \"cluster-image-registry-operator-dc59b4c8b-2jkt6\" (UID: \"7a17670c-87d6-4fc7-b197-6817f4467377\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747865 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sb4p\" (UniqueName: \"kubernetes.io/projected/df33868c-7e7c-4d67-b478-40ee1d7ef69d-kube-api-access-2sb4p\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747896 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-audit-policies\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747917 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq54s\" (UniqueName: \"kubernetes.io/projected/9af7072b-a939-4cdb-813f-1dccb589763a-kube-api-access-sq54s\") pod \"cluster-samples-operator-665b6dd947-9qz4w\" (UID: \"9af7072b-a939-4cdb-813f-1dccb589763a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9qz4w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747940 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f7d5d65a-b6dd-45dd-9f35-5e02a915a64f-stats-auth\") pod \"router-default-5444994796-nr4zg\" (UID: \"f7d5d65a-b6dd-45dd-9f35-5e02a915a64f\") " pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747959 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747980 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltpgl\" (UniqueName: \"kubernetes.io/projected/fbb37aa9-5a46-4ef3-9251-748a657e0933-kube-api-access-ltpgl\") pod \"openshift-controller-manager-operator-756b6f6bc6-zpl2d\" (UID: \"fbb37aa9-5a46-4ef3-9251-748a657e0933\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zpl2d" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.748000 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/df33868c-7e7c-4d67-b478-40ee1d7ef69d-console-serving-cert\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.748032 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/13d7ba8a-bacc-4294-8e56-c3425f284c1a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sl8lb\" (UID: \"13d7ba8a-bacc-4294-8e56-c3425f284c1a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sl8lb" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.748043 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0-available-featuregates\") pod \"openshift-config-operator-7777fb866f-smpjk\" (UID: \"b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.748056 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91ee6068-d199-4b76-81fa-897155d5a38d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7mpcc\" (UID: \"91ee6068-d199-4b76-81fa-897155d5a38d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mpcc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.748716 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/df33868c-7e7c-4d67-b478-40ee1d7ef69d-oauth-serving-cert\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.748769 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/63972ba9-04aa-42e8-a91c-f27796ab40f4-audit-dir\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.748916 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/df33868c-7e7c-4d67-b478-40ee1d7ef69d-console-config\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.749186 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62710ae2-1843-4880-8444-b501a206e145-config\") pod \"authentication-operator-69f744f599-2prtt\" (UID: \"62710ae2-1843-4880-8444-b501a206e145\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.750286 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f7d5d65a-b6dd-45dd-9f35-5e02a915a64f-metrics-certs\") pod \"router-default-5444994796-nr4zg\" (UID: \"f7d5d65a-b6dd-45dd-9f35-5e02a915a64f\") " pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.750872 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.750900 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.750936 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62710ae2-1843-4880-8444-b501a206e145-serving-cert\") pod \"authentication-operator-69f744f599-2prtt\" (UID: \"62710ae2-1843-4880-8444-b501a206e145\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.747960 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7d5d65a-b6dd-45dd-9f35-5e02a915a64f-service-ca-bundle\") pod \"router-default-5444994796-nr4zg\" (UID: \"f7d5d65a-b6dd-45dd-9f35-5e02a915a64f\") " pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.751042 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f7d5d65a-b6dd-45dd-9f35-5e02a915a64f-default-certificate\") pod \"router-default-5444994796-nr4zg\" (UID: \"f7d5d65a-b6dd-45dd-9f35-5e02a915a64f\") " pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.751566 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df33868c-7e7c-4d67-b478-40ee1d7ef69d-trusted-ca-bundle\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.751734 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.751954 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/df33868c-7e7c-4d67-b478-40ee1d7ef69d-service-ca\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.751955 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.751983 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-audit-policies\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.752259 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9af7072b-a939-4cdb-813f-1dccb589763a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9qz4w\" (UID: \"9af7072b-a939-4cdb-813f-1dccb589763a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9qz4w" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.752662 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91ee6068-d199-4b76-81fa-897155d5a38d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7mpcc\" (UID: \"91ee6068-d199-4b76-81fa-897155d5a38d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mpcc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.752705 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62710ae2-1843-4880-8444-b501a206e145-service-ca-bundle\") pod \"authentication-operator-69f744f599-2prtt\" (UID: \"62710ae2-1843-4880-8444-b501a206e145\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.753407 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0-serving-cert\") pod \"openshift-config-operator-7777fb866f-smpjk\" (UID: \"b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.753684 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.753913 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.754427 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.754440 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7a17670c-87d6-4fc7-b197-6817f4467377-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2jkt6\" (UID: \"7a17670c-87d6-4fc7-b197-6817f4467377\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.754932 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91ee6068-d199-4b76-81fa-897155d5a38d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7mpcc\" (UID: \"91ee6068-d199-4b76-81fa-897155d5a38d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mpcc" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.754938 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/7a17670c-87d6-4fc7-b197-6817f4467377-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2jkt6\" (UID: \"7a17670c-87d6-4fc7-b197-6817f4467377\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.755091 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/df33868c-7e7c-4d67-b478-40ee1d7ef69d-console-serving-cert\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.755777 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.756003 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.756781 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.758481 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.758486 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f7d5d65a-b6dd-45dd-9f35-5e02a915a64f-stats-auth\") pod \"router-default-5444994796-nr4zg\" (UID: \"f7d5d65a-b6dd-45dd-9f35-5e02a915a64f\") " pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.770331 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/df33868c-7e7c-4d67-b478-40ee1d7ef69d-console-oauth-config\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.775202 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.794712 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.815266 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.835088 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.848722 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/17422b05-1782-421b-ab25-4ce61d267fc2-config-volume\") pod \"collect-profiles-29401485-28zdh\" (UID: \"17422b05-1782-421b-ab25-4ce61d267fc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.848965 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/17422b05-1782-421b-ab25-4ce61d267fc2-secret-volume\") pod \"collect-profiles-29401485-28zdh\" (UID: \"17422b05-1782-421b-ab25-4ce61d267fc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.849142 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2mcn\" (UniqueName: \"kubernetes.io/projected/8a3ebc4f-0491-4dc1-8c5d-adb851b6ef87-kube-api-access-w2mcn\") pod \"downloads-7954f5f757-z8dgw\" (UID: \"8a3ebc4f-0491-4dc1-8c5d-adb851b6ef87\") " pod="openshift-console/downloads-7954f5f757-z8dgw" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.849273 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t96bv\" (UniqueName: \"kubernetes.io/projected/17422b05-1782-421b-ab25-4ce61d267fc2-kube-api-access-t96bv\") pod \"collect-profiles-29401485-28zdh\" (UID: \"17422b05-1782-421b-ab25-4ce61d267fc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.849557 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbb37aa9-5a46-4ef3-9251-748a657e0933-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zpl2d\" (UID: \"fbb37aa9-5a46-4ef3-9251-748a657e0933\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zpl2d" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.849672 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13d7ba8a-bacc-4294-8e56-c3425f284c1a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sl8lb\" (UID: \"13d7ba8a-bacc-4294-8e56-c3425f284c1a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sl8lb" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.849826 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltpgl\" (UniqueName: \"kubernetes.io/projected/fbb37aa9-5a46-4ef3-9251-748a657e0933-kube-api-access-ltpgl\") pod \"openshift-controller-manager-operator-756b6f6bc6-zpl2d\" (UID: \"fbb37aa9-5a46-4ef3-9251-748a657e0933\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zpl2d" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.849949 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/13d7ba8a-bacc-4294-8e56-c3425f284c1a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sl8lb\" (UID: \"13d7ba8a-bacc-4294-8e56-c3425f284c1a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sl8lb" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.850078 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fbb37aa9-5a46-4ef3-9251-748a657e0933-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zpl2d\" (UID: \"fbb37aa9-5a46-4ef3-9251-748a657e0933\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zpl2d" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.850185 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13d7ba8a-bacc-4294-8e56-c3425f284c1a-config\") pod \"kube-controller-manager-operator-78b949d7b-sl8lb\" (UID: \"13d7ba8a-bacc-4294-8e56-c3425f284c1a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sl8lb" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.855176 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.876614 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.895989 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.915409 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.935110 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.956418 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.976488 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 16:49:15 crc kubenswrapper[4802]: I1125 16:49:15.995342 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.015978 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.025376 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fbb37aa9-5a46-4ef3-9251-748a657e0933-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zpl2d\" (UID: \"fbb37aa9-5a46-4ef3-9251-748a657e0933\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zpl2d" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.035397 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.055557 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.060363 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbb37aa9-5a46-4ef3-9251-748a657e0933-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zpl2d\" (UID: \"fbb37aa9-5a46-4ef3-9251-748a657e0933\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zpl2d" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.075698 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.095272 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.115806 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.135426 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.175223 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.195769 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.215685 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.223479 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13d7ba8a-bacc-4294-8e56-c3425f284c1a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sl8lb\" (UID: \"13d7ba8a-bacc-4294-8e56-c3425f284c1a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sl8lb" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.234907 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.241283 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13d7ba8a-bacc-4294-8e56-c3425f284c1a-config\") pod \"kube-controller-manager-operator-78b949d7b-sl8lb\" (UID: \"13d7ba8a-bacc-4294-8e56-c3425f284c1a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sl8lb" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.255592 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.276968 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.296501 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.315549 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.335238 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.355283 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.374878 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.394768 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.415335 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.435867 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.455380 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.476014 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.496181 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.515502 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.535088 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.556025 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.576117 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.594363 4802 request.go:700] Waited for 1.000668234s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console-operator/configmaps?fieldSelector=metadata.name%3Dtrusted-ca&limit=500&resourceVersion=0 Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.601384 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.615842 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.634918 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.656284 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.676324 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.696435 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.716301 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.735638 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.756398 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.763872 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/17422b05-1782-421b-ab25-4ce61d267fc2-secret-volume\") pod \"collect-profiles-29401485-28zdh\" (UID: \"17422b05-1782-421b-ab25-4ce61d267fc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.775145 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.795204 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.816113 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.836116 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 16:49:16 crc kubenswrapper[4802]: E1125 16:49:16.849061 4802 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Nov 25 16:49:16 crc kubenswrapper[4802]: E1125 16:49:16.849238 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/17422b05-1782-421b-ab25-4ce61d267fc2-config-volume podName:17422b05-1782-421b-ab25-4ce61d267fc2 nodeName:}" failed. No retries permitted until 2025-11-25 16:49:17.349208424 +0000 UTC m=+140.493555610 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/17422b05-1782-421b-ab25-4ce61d267fc2-config-volume") pod "collect-profiles-29401485-28zdh" (UID: "17422b05-1782-421b-ab25-4ce61d267fc2") : failed to sync configmap cache: timed out waiting for the condition Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.855993 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.876688 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.896605 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.915615 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.935884 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.955888 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.976580 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 16:49:16 crc kubenswrapper[4802]: I1125 16:49:16.996958 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.016244 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.036110 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.056106 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.077031 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.096314 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.115826 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.135050 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.155904 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.176203 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.202358 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.216013 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.235599 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.255452 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.275988 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.295908 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.316412 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.335780 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.372604 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n8jz\" (UniqueName: \"kubernetes.io/projected/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-kube-api-access-4n8jz\") pod \"route-controller-manager-6576b87f9c-zvb9w\" (UID: \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.372857 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/17422b05-1782-421b-ab25-4ce61d267fc2-config-volume\") pod \"collect-profiles-29401485-28zdh\" (UID: \"17422b05-1782-421b-ab25-4ce61d267fc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.373635 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/17422b05-1782-421b-ab25-4ce61d267fc2-config-volume\") pod \"collect-profiles-29401485-28zdh\" (UID: \"17422b05-1782-421b-ab25-4ce61d267fc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.391403 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4x99\" (UniqueName: \"kubernetes.io/projected/49937863-524b-4ee0-b3c4-6522995947c4-kube-api-access-t4x99\") pod \"controller-manager-879f6c89f-bv6b6\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.411226 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7ljt\" (UniqueName: \"kubernetes.io/projected/c673008d-695e-44a5-8b6d-1d7c585b5b72-kube-api-access-n7ljt\") pod \"machine-approver-56656f9798-2wcfc\" (UID: \"c673008d-695e-44a5-8b6d-1d7c585b5b72\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.416925 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.451880 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwl2t\" (UniqueName: \"kubernetes.io/projected/54c5e683-04bc-4241-8d13-17b415237707-kube-api-access-zwl2t\") pod \"apiserver-7bbb656c7d-hljn2\" (UID: \"54c5e683-04bc-4241-8d13-17b415237707\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.470934 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v57bl\" (UniqueName: \"kubernetes.io/projected/ea30ecf0-8806-4c5a-9ff6-2738ca72e50b-kube-api-access-v57bl\") pod \"machine-api-operator-5694c8668f-j2t9z\" (UID: \"ea30ecf0-8806-4c5a-9ff6-2738ca72e50b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j2t9z" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.490462 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z46ch\" (UniqueName: \"kubernetes.io/projected/a0a85d28-62a0-4ddc-ab49-c73b6db760a0-kube-api-access-z46ch\") pod \"apiserver-76f77b778f-hzx2q\" (UID: \"a0a85d28-62a0-4ddc-ab49-c73b6db760a0\") " pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.515599 4802 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.534992 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.549511 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.556283 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.563889 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.577975 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.581328 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.591697 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.594000 4802 request.go:700] Waited for 1.886239709s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dnode-bootstrapper-token&limit=500&resourceVersion=0 Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.597052 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 16:49:17 crc kubenswrapper[4802]: W1125 16:49:17.599542 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc673008d_695e_44a5_8b6d_1d7c585b5b72.slice/crio-4071992d60e25853469b65e592fe3b986195561358622331170db88c90442e4b WatchSource:0}: Error finding container 4071992d60e25853469b65e592fe3b986195561358622331170db88c90442e4b: Status 404 returned error can't find the container with id 4071992d60e25853469b65e592fe3b986195561358622331170db88c90442e4b Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.615628 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.626417 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-j2t9z" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.648318 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.652374 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7a17670c-87d6-4fc7-b197-6817f4467377-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2jkt6\" (UID: \"7a17670c-87d6-4fc7-b197-6817f4467377\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.687461 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqwm7\" (UniqueName: \"kubernetes.io/projected/62710ae2-1843-4880-8444-b501a206e145-kube-api-access-fqwm7\") pod \"authentication-operator-69f744f599-2prtt\" (UID: \"62710ae2-1843-4880-8444-b501a206e145\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.694345 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k25r\" (UniqueName: \"kubernetes.io/projected/91ee6068-d199-4b76-81fa-897155d5a38d-kube-api-access-6k25r\") pod \"openshift-apiserver-operator-796bbdcf4f-7mpcc\" (UID: \"91ee6068-d199-4b76-81fa-897155d5a38d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mpcc" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.712409 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.723222 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfvrl\" (UniqueName: \"kubernetes.io/projected/b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0-kube-api-access-qfvrl\") pod \"openshift-config-operator-7777fb866f-smpjk\" (UID: \"b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.734487 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm5mh\" (UniqueName: \"kubernetes.io/projected/f7d5d65a-b6dd-45dd-9f35-5e02a915a64f-kube-api-access-lm5mh\") pod \"router-default-5444994796-nr4zg\" (UID: \"f7d5d65a-b6dd-45dd-9f35-5e02a915a64f\") " pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.743560 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.758096 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sb4p\" (UniqueName: \"kubernetes.io/projected/df33868c-7e7c-4d67-b478-40ee1d7ef69d-kube-api-access-2sb4p\") pod \"console-f9d7485db-p7zvz\" (UID: \"df33868c-7e7c-4d67-b478-40ee1d7ef69d\") " pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.778390 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w9gv\" (UniqueName: \"kubernetes.io/projected/7a17670c-87d6-4fc7-b197-6817f4467377-kube-api-access-8w9gv\") pod \"cluster-image-registry-operator-dc59b4c8b-2jkt6\" (UID: \"7a17670c-87d6-4fc7-b197-6817f4467377\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.796190 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq54s\" (UniqueName: \"kubernetes.io/projected/9af7072b-a939-4cdb-813f-1dccb589763a-kube-api-access-sq54s\") pod \"cluster-samples-operator-665b6dd947-9qz4w\" (UID: \"9af7072b-a939-4cdb-813f-1dccb589763a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9qz4w" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.804935 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.810533 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.814541 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8npcc\" (UniqueName: \"kubernetes.io/projected/63972ba9-04aa-42e8-a91c-f27796ab40f4-kube-api-access-8npcc\") pod \"oauth-openshift-558db77b4-sccr2\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.835273 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2mcn\" (UniqueName: \"kubernetes.io/projected/8a3ebc4f-0491-4dc1-8c5d-adb851b6ef87-kube-api-access-w2mcn\") pod \"downloads-7954f5f757-z8dgw\" (UID: \"8a3ebc4f-0491-4dc1-8c5d-adb851b6ef87\") " pod="openshift-console/downloads-7954f5f757-z8dgw" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.850402 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t96bv\" (UniqueName: \"kubernetes.io/projected/17422b05-1782-421b-ab25-4ce61d267fc2-kube-api-access-t96bv\") pod \"collect-profiles-29401485-28zdh\" (UID: \"17422b05-1782-421b-ab25-4ce61d267fc2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.872692 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltpgl\" (UniqueName: \"kubernetes.io/projected/fbb37aa9-5a46-4ef3-9251-748a657e0933-kube-api-access-ltpgl\") pod \"openshift-controller-manager-operator-756b6f6bc6-zpl2d\" (UID: \"fbb37aa9-5a46-4ef3-9251-748a657e0933\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zpl2d" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.907864 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/13d7ba8a-bacc-4294-8e56-c3425f284c1a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sl8lb\" (UID: \"13d7ba8a-bacc-4294-8e56-c3425f284c1a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sl8lb" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.912931 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j2t9z"] Nov 25 16:49:17 crc kubenswrapper[4802]: W1125 16:49:17.928740 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea30ecf0_8806_4c5a_9ff6_2738ca72e50b.slice/crio-c18bc0412072957ec471138eabefc800b6c681265e24d417b182604fa91a83b7 WatchSource:0}: Error finding container c18bc0412072957ec471138eabefc800b6c681265e24d417b182604fa91a83b7: Status 404 returned error can't find the container with id c18bc0412072957ec471138eabefc800b6c681265e24d417b182604fa91a83b7 Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.946236 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-z8dgw" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.958906 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mpcc" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.974832 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.979772 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983319 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fc06abdb-2b20-4fbd-8766-7470b97b3e27-trusted-ca\") pod \"ingress-operator-5b745b69d9-k2xl9\" (UID: \"fc06abdb-2b20-4fbd-8766-7470b97b3e27\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983387 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khkxd\" (UniqueName: \"kubernetes.io/projected/7583c399-65b2-4e40-91eb-1908753b6a29-kube-api-access-khkxd\") pod \"control-plane-machine-set-operator-78cbb6b69f-9lzl9\" (UID: \"7583c399-65b2-4e40-91eb-1908753b6a29\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9lzl9" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983416 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aad8389e-af3c-4783-bf7d-12945fc25070-serving-cert\") pod \"console-operator-58897d9998-nmzxv\" (UID: \"aad8389e-af3c-4783-bf7d-12945fc25070\") " pod="openshift-console-operator/console-operator-58897d9998-nmzxv" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983440 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1f572606-6161-4ac4-a33a-8adaea8c071a-signing-cabundle\") pod \"service-ca-9c57cc56f-ppfnt\" (UID: \"1f572606-6161-4ac4-a33a-8adaea8c071a\") " pod="openshift-service-ca/service-ca-9c57cc56f-ppfnt" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983461 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a9c03eb-ba03-4b29-b3ad-585f9873973b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dzmm9\" (UID: \"8a9c03eb-ba03-4b29-b3ad-585f9873973b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmm9" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983508 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/030b390b-1901-4a54-98f8-bf241a955c22-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jzd75\" (UID: \"030b390b-1901-4a54-98f8-bf241a955c22\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jzd75" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983530 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a9c03eb-ba03-4b29-b3ad-585f9873973b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dzmm9\" (UID: \"8a9c03eb-ba03-4b29-b3ad-585f9873973b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmm9" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983565 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/030b390b-1901-4a54-98f8-bf241a955c22-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jzd75\" (UID: \"030b390b-1901-4a54-98f8-bf241a955c22\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jzd75" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983613 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c5d984b-9b31-4997-904a-f27989f78a8c-bound-sa-token\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983636 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qkx8\" (UniqueName: \"kubernetes.io/projected/fc06abdb-2b20-4fbd-8766-7470b97b3e27-kube-api-access-4qkx8\") pod \"ingress-operator-5b745b69d9-k2xl9\" (UID: \"fc06abdb-2b20-4fbd-8766-7470b97b3e27\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983659 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/88c056b9-7ee2-40be-b5ec-abec9f736e3c-srv-cert\") pod \"olm-operator-6b444d44fb-29vlk\" (UID: \"88c056b9-7ee2-40be-b5ec-abec9f736e3c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983682 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/392432cc-4ff4-4466-acac-7298b2a6197b-tmpfs\") pod \"packageserver-d55dfcdfc-hb9lv\" (UID: \"392432cc-4ff4-4466-acac-7298b2a6197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983718 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/030b390b-1901-4a54-98f8-bf241a955c22-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jzd75\" (UID: \"030b390b-1901-4a54-98f8-bf241a955c22\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jzd75" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983754 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d3ecea3d-c41f-46f4-bf0d-d43156542fa6-images\") pod \"machine-config-operator-74547568cd-c8xrm\" (UID: \"d3ecea3d-c41f-46f4-bf0d-d43156542fa6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983777 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2852a9e-0047-4ae1-941d-dde0a4cc991f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-zjrnl\" (UID: \"c2852a9e-0047-4ae1-941d-dde0a4cc991f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zjrnl" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983800 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9tt57\" (UID: \"042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9tt57" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983822 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/392432cc-4ff4-4466-acac-7298b2a6197b-apiservice-cert\") pod \"packageserver-d55dfcdfc-hb9lv\" (UID: \"392432cc-4ff4-4466-acac-7298b2a6197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983844 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e73052a1-e668-48c1-8734-5a3dc1f6b5f2-metrics-tls\") pod \"dns-default-v6gj4\" (UID: \"e73052a1-e668-48c1-8734-5a3dc1f6b5f2\") " pod="openshift-dns/dns-default-v6gj4" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983897 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d3ecea3d-c41f-46f4-bf0d-d43156542fa6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-c8xrm\" (UID: \"d3ecea3d-c41f-46f4-bf0d-d43156542fa6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983921 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9-config\") pod \"kube-apiserver-operator-766d6c64bb-9tt57\" (UID: \"042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9tt57" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983940 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aad8389e-af3c-4783-bf7d-12945fc25070-trusted-ca\") pod \"console-operator-58897d9998-nmzxv\" (UID: \"aad8389e-af3c-4783-bf7d-12945fc25070\") " pod="openshift-console-operator/console-operator-58897d9998-nmzxv" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983962 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1f572606-6161-4ac4-a33a-8adaea8c071a-signing-key\") pod \"service-ca-9c57cc56f-ppfnt\" (UID: \"1f572606-6161-4ac4-a33a-8adaea8c071a\") " pod="openshift-service-ca/service-ca-9c57cc56f-ppfnt" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.983984 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9tt57\" (UID: \"042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9tt57" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984006 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aad8389e-af3c-4783-bf7d-12945fc25070-config\") pod \"console-operator-58897d9998-nmzxv\" (UID: \"aad8389e-af3c-4783-bf7d-12945fc25070\") " pod="openshift-console-operator/console-operator-58897d9998-nmzxv" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984062 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdn88\" (UniqueName: \"kubernetes.io/projected/1f572606-6161-4ac4-a33a-8adaea8c071a-kube-api-access-wdn88\") pod \"service-ca-9c57cc56f-ppfnt\" (UID: \"1f572606-6161-4ac4-a33a-8adaea8c071a\") " pod="openshift-service-ca/service-ca-9c57cc56f-ppfnt" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984082 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b09c8b8b-8696-4a22-8b9a-4125ab33bc33-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-852rn\" (UID: \"b09c8b8b-8696-4a22-8b9a-4125ab33bc33\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-852rn" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984103 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sgsx\" (UniqueName: \"kubernetes.io/projected/fd37087b-7ebe-4f7f-9121-d55c058d804b-kube-api-access-6sgsx\") pod \"dns-operator-744455d44c-79vhf\" (UID: \"fd37087b-7ebe-4f7f-9121-d55c058d804b\") " pod="openshift-dns-operator/dns-operator-744455d44c-79vhf" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984157 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64qpx\" (UniqueName: \"kubernetes.io/projected/88c056b9-7ee2-40be-b5ec-abec9f736e3c-kube-api-access-64qpx\") pod \"olm-operator-6b444d44fb-29vlk\" (UID: \"88c056b9-7ee2-40be-b5ec-abec9f736e3c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984180 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d33f1ce9-b084-4ef3-887b-84f2d984f2a8-etcd-client\") pod \"etcd-operator-b45778765-nh2kl\" (UID: \"d33f1ce9-b084-4ef3-887b-84f2d984f2a8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984212 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzp7w\" (UniqueName: \"kubernetes.io/projected/c2852a9e-0047-4ae1-941d-dde0a4cc991f-kube-api-access-nzp7w\") pod \"package-server-manager-789f6589d5-zjrnl\" (UID: \"c2852a9e-0047-4ae1-941d-dde0a4cc991f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zjrnl" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984244 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2d266390-de7a-4456-ae91-f44fe6a8b532-srv-cert\") pod \"catalog-operator-68c6474976-rlnls\" (UID: \"2d266390-de7a-4456-ae91-f44fe6a8b532\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984310 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wnkh\" (UniqueName: \"kubernetes.io/projected/8a9c03eb-ba03-4b29-b3ad-585f9873973b-kube-api-access-6wnkh\") pod \"kube-storage-version-migrator-operator-b67b599dd-dzmm9\" (UID: \"8a9c03eb-ba03-4b29-b3ad-585f9873973b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmm9" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984344 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mv96\" (UniqueName: \"kubernetes.io/projected/d33f1ce9-b084-4ef3-887b-84f2d984f2a8-kube-api-access-4mv96\") pod \"etcd-operator-b45778765-nh2kl\" (UID: \"d33f1ce9-b084-4ef3-887b-84f2d984f2a8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984371 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t22ct\" (UniqueName: \"kubernetes.io/projected/d3ecea3d-c41f-46f4-bf0d-d43156542fa6-kube-api-access-t22ct\") pod \"machine-config-operator-74547568cd-c8xrm\" (UID: \"d3ecea3d-c41f-46f4-bf0d-d43156542fa6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984395 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt9qm\" (UniqueName: \"kubernetes.io/projected/611d3c60-91d0-45f5-a3aa-dffcc427ef0b-kube-api-access-kt9qm\") pod \"service-ca-operator-777779d784-6hwbq\" (UID: \"611d3c60-91d0-45f5-a3aa-dffcc427ef0b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6hwbq" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984436 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st6fd\" (UniqueName: \"kubernetes.io/projected/aad8389e-af3c-4783-bf7d-12945fc25070-kube-api-access-st6fd\") pod \"console-operator-58897d9998-nmzxv\" (UID: \"aad8389e-af3c-4783-bf7d-12945fc25070\") " pod="openshift-console-operator/console-operator-58897d9998-nmzxv" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984470 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28bbp\" (UniqueName: \"kubernetes.io/projected/b09c8b8b-8696-4a22-8b9a-4125ab33bc33-kube-api-access-28bbp\") pod \"machine-config-controller-84d6567774-852rn\" (UID: \"b09c8b8b-8696-4a22-8b9a-4125ab33bc33\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-852rn" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984487 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt8nb\" (UniqueName: \"kubernetes.io/projected/e73052a1-e668-48c1-8734-5a3dc1f6b5f2-kube-api-access-rt8nb\") pod \"dns-default-v6gj4\" (UID: \"e73052a1-e668-48c1-8734-5a3dc1f6b5f2\") " pod="openshift-dns/dns-default-v6gj4" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984516 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d33f1ce9-b084-4ef3-887b-84f2d984f2a8-serving-cert\") pod \"etcd-operator-b45778765-nh2kl\" (UID: \"d33f1ce9-b084-4ef3-887b-84f2d984f2a8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984553 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d3ecea3d-c41f-46f4-bf0d-d43156542fa6-proxy-tls\") pod \"machine-config-operator-74547568cd-c8xrm\" (UID: \"d3ecea3d-c41f-46f4-bf0d-d43156542fa6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984582 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnqsr\" (UniqueName: \"kubernetes.io/projected/392432cc-4ff4-4466-acac-7298b2a6197b-kube-api-access-nnqsr\") pod \"packageserver-d55dfcdfc-hb9lv\" (UID: \"392432cc-4ff4-4466-acac-7298b2a6197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984664 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fc06abdb-2b20-4fbd-8766-7470b97b3e27-metrics-tls\") pod \"ingress-operator-5b745b69d9-k2xl9\" (UID: \"fc06abdb-2b20-4fbd-8766-7470b97b3e27\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984693 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2d266390-de7a-4456-ae91-f44fe6a8b532-profile-collector-cert\") pod \"catalog-operator-68c6474976-rlnls\" (UID: \"2d266390-de7a-4456-ae91-f44fe6a8b532\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984712 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8c5d984b-9b31-4997-904a-f27989f78a8c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984729 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/611d3c60-91d0-45f5-a3aa-dffcc427ef0b-config\") pod \"service-ca-operator-777779d784-6hwbq\" (UID: \"611d3c60-91d0-45f5-a3aa-dffcc427ef0b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6hwbq" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984746 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d33f1ce9-b084-4ef3-887b-84f2d984f2a8-config\") pod \"etcd-operator-b45778765-nh2kl\" (UID: \"d33f1ce9-b084-4ef3-887b-84f2d984f2a8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984761 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d33f1ce9-b084-4ef3-887b-84f2d984f2a8-etcd-ca\") pod \"etcd-operator-b45778765-nh2kl\" (UID: \"d33f1ce9-b084-4ef3-887b-84f2d984f2a8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984779 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f5d5515f-a284-4278-8c00-9222ee11b428-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-hk4bx\" (UID: \"f5d5515f-a284-4278-8c00-9222ee11b428\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hk4bx" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984798 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98e0e18f-d88b-416a-b18b-eaaea13c8e21-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cr9nk\" (UID: \"98e0e18f-d88b-416a-b18b-eaaea13c8e21\") " pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984817 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/88c056b9-7ee2-40be-b5ec-abec9f736e3c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-29vlk\" (UID: \"88c056b9-7ee2-40be-b5ec-abec9f736e3c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984834 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8c5d984b-9b31-4997-904a-f27989f78a8c-registry-certificates\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984850 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fd37087b-7ebe-4f7f-9121-d55c058d804b-metrics-tls\") pod \"dns-operator-744455d44c-79vhf\" (UID: \"fd37087b-7ebe-4f7f-9121-d55c058d804b\") " pod="openshift-dns-operator/dns-operator-744455d44c-79vhf" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984865 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqqmg\" (UniqueName: \"kubernetes.io/projected/ac950124-b8dd-4603-a82e-6e2484ce2aaa-kube-api-access-rqqmg\") pod \"migrator-59844c95c7-kk8nf\" (UID: \"ac950124-b8dd-4603-a82e-6e2484ce2aaa\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kk8nf" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984893 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8c5d984b-9b31-4997-904a-f27989f78a8c-registry-tls\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.984947 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7583c399-65b2-4e40-91eb-1908753b6a29-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9lzl9\" (UID: \"7583c399-65b2-4e40-91eb-1908753b6a29\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9lzl9" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.985339 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgnl2\" (UniqueName: \"kubernetes.io/projected/8c5d984b-9b31-4997-904a-f27989f78a8c-kube-api-access-dgnl2\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.985364 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/06d8bfc5-8156-4181-845b-dde35a4be1fe-cert\") pod \"ingress-canary-cfw4c\" (UID: \"06d8bfc5-8156-4181-845b-dde35a4be1fe\") " pod="openshift-ingress-canary/ingress-canary-cfw4c" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.985410 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzb85\" (UniqueName: \"kubernetes.io/projected/f5d5515f-a284-4278-8c00-9222ee11b428-kube-api-access-hzb85\") pod \"multus-admission-controller-857f4d67dd-hk4bx\" (UID: \"f5d5515f-a284-4278-8c00-9222ee11b428\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hk4bx" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.985439 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c5d984b-9b31-4997-904a-f27989f78a8c-trusted-ca\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.985461 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w29tv\" (UniqueName: \"kubernetes.io/projected/98e0e18f-d88b-416a-b18b-eaaea13c8e21-kube-api-access-w29tv\") pod \"marketplace-operator-79b997595-cr9nk\" (UID: \"98e0e18f-d88b-416a-b18b-eaaea13c8e21\") " pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.985483 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l85qq\" (UniqueName: \"kubernetes.io/projected/06d8bfc5-8156-4181-845b-dde35a4be1fe-kube-api-access-l85qq\") pod \"ingress-canary-cfw4c\" (UID: \"06d8bfc5-8156-4181-845b-dde35a4be1fe\") " pod="openshift-ingress-canary/ingress-canary-cfw4c" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.985498 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d33f1ce9-b084-4ef3-887b-84f2d984f2a8-etcd-service-ca\") pod \"etcd-operator-b45778765-nh2kl\" (UID: \"d33f1ce9-b084-4ef3-887b-84f2d984f2a8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.985534 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.985554 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fc06abdb-2b20-4fbd-8766-7470b97b3e27-bound-sa-token\") pod \"ingress-operator-5b745b69d9-k2xl9\" (UID: \"fc06abdb-2b20-4fbd-8766-7470b97b3e27\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.985587 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8c5d984b-9b31-4997-904a-f27989f78a8c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.985609 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e73052a1-e668-48c1-8734-5a3dc1f6b5f2-config-volume\") pod \"dns-default-v6gj4\" (UID: \"e73052a1-e668-48c1-8734-5a3dc1f6b5f2\") " pod="openshift-dns/dns-default-v6gj4" Nov 25 16:49:17 crc kubenswrapper[4802]: E1125 16:49:17.986271 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:18.486257791 +0000 UTC m=+141.630604977 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.986344 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b09c8b8b-8696-4a22-8b9a-4125ab33bc33-proxy-tls\") pod \"machine-config-controller-84d6567774-852rn\" (UID: \"b09c8b8b-8696-4a22-8b9a-4125ab33bc33\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-852rn" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.986376 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/98e0e18f-d88b-416a-b18b-eaaea13c8e21-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cr9nk\" (UID: \"98e0e18f-d88b-416a-b18b-eaaea13c8e21\") " pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.986600 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjnrf\" (UniqueName: \"kubernetes.io/projected/2d266390-de7a-4456-ae91-f44fe6a8b532-kube-api-access-mjnrf\") pod \"catalog-operator-68c6474976-rlnls\" (UID: \"2d266390-de7a-4456-ae91-f44fe6a8b532\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.986683 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/611d3c60-91d0-45f5-a3aa-dffcc427ef0b-serving-cert\") pod \"service-ca-operator-777779d784-6hwbq\" (UID: \"611d3c60-91d0-45f5-a3aa-dffcc427ef0b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6hwbq" Nov 25 16:49:17 crc kubenswrapper[4802]: I1125 16:49:17.986793 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/392432cc-4ff4-4466-acac-7298b2a6197b-webhook-cert\") pod \"packageserver-d55dfcdfc-hb9lv\" (UID: \"392432cc-4ff4-4466-acac-7298b2a6197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.003023 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.005926 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w"] Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.024614 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9qz4w" Nov 25 16:49:18 crc kubenswrapper[4802]: W1125 16:49:18.026764 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef963c3e_4e73_4fe3_aed1_21d8b7d8be40.slice/crio-209d464547d064d8f2f3c862d07b0554d0f34b6a162ba999435821ef00f67dc3 WatchSource:0}: Error finding container 209d464547d064d8f2f3c862d07b0554d0f34b6a162ba999435821ef00f67dc3: Status 404 returned error can't find the container with id 209d464547d064d8f2f3c862d07b0554d0f34b6a162ba999435821ef00f67dc3 Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.028544 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2"] Nov 25 16:49:18 crc kubenswrapper[4802]: W1125 16:49:18.047043 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54c5e683_04bc_4241_8d13_17b415237707.slice/crio-ca6be1fa1059914add7f8891845767b0f3a87aca14f8a37d1225684056613721 WatchSource:0}: Error finding container ca6be1fa1059914add7f8891845767b0f3a87aca14f8a37d1225684056613721: Status 404 returned error can't find the container with id ca6be1fa1059914add7f8891845767b0f3a87aca14f8a37d1225684056613721 Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.067860 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" event={"ID":"54c5e683-04bc-4241-8d13-17b415237707","Type":"ContainerStarted","Data":"ca6be1fa1059914add7f8891845767b0f3a87aca14f8a37d1225684056613721"} Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.068907 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" event={"ID":"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40","Type":"ContainerStarted","Data":"209d464547d064d8f2f3c862d07b0554d0f34b6a162ba999435821ef00f67dc3"} Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.069836 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc" event={"ID":"c673008d-695e-44a5-8b6d-1d7c585b5b72","Type":"ContainerStarted","Data":"4071992d60e25853469b65e592fe3b986195561358622331170db88c90442e4b"} Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.071115 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-nr4zg" event={"ID":"f7d5d65a-b6dd-45dd-9f35-5e02a915a64f","Type":"ContainerStarted","Data":"c1c9c548b04d63af6ced4845c161da45b6e9a836f5860997907bb63972da801b"} Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.072418 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j2t9z" event={"ID":"ea30ecf0-8806-4c5a-9ff6-2738ca72e50b","Type":"ContainerStarted","Data":"c18bc0412072957ec471138eabefc800b6c681265e24d417b182604fa91a83b7"} Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.087573 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.087789 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d27bd203-adaf-4959-8907-416cd199df98-registration-dir\") pod \"csi-hostpathplugin-s684g\" (UID: \"d27bd203-adaf-4959-8907-416cd199df98\") " pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: E1125 16:49:18.087846 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:18.587813691 +0000 UTC m=+141.732160877 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.087970 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d33f1ce9-b084-4ef3-887b-84f2d984f2a8-serving-cert\") pod \"etcd-operator-b45778765-nh2kl\" (UID: \"d33f1ce9-b084-4ef3-887b-84f2d984f2a8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088008 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d3ecea3d-c41f-46f4-bf0d-d43156542fa6-proxy-tls\") pod \"machine-config-operator-74547568cd-c8xrm\" (UID: \"d3ecea3d-c41f-46f4-bf0d-d43156542fa6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088038 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnqsr\" (UniqueName: \"kubernetes.io/projected/392432cc-4ff4-4466-acac-7298b2a6197b-kube-api-access-nnqsr\") pod \"packageserver-d55dfcdfc-hb9lv\" (UID: \"392432cc-4ff4-4466-acac-7298b2a6197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088061 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52k5m\" (UniqueName: \"kubernetes.io/projected/7635f2d4-3be1-490a-99b9-3783a3fc58a4-kube-api-access-52k5m\") pod \"machine-config-server-4sl2t\" (UID: \"7635f2d4-3be1-490a-99b9-3783a3fc58a4\") " pod="openshift-machine-config-operator/machine-config-server-4sl2t" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088132 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fc06abdb-2b20-4fbd-8766-7470b97b3e27-metrics-tls\") pod \"ingress-operator-5b745b69d9-k2xl9\" (UID: \"fc06abdb-2b20-4fbd-8766-7470b97b3e27\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088153 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8c5d984b-9b31-4997-904a-f27989f78a8c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088168 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2d266390-de7a-4456-ae91-f44fe6a8b532-profile-collector-cert\") pod \"catalog-operator-68c6474976-rlnls\" (UID: \"2d266390-de7a-4456-ae91-f44fe6a8b532\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088183 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f5d5515f-a284-4278-8c00-9222ee11b428-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-hk4bx\" (UID: \"f5d5515f-a284-4278-8c00-9222ee11b428\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hk4bx" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088199 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/611d3c60-91d0-45f5-a3aa-dffcc427ef0b-config\") pod \"service-ca-operator-777779d784-6hwbq\" (UID: \"611d3c60-91d0-45f5-a3aa-dffcc427ef0b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6hwbq" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088223 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d33f1ce9-b084-4ef3-887b-84f2d984f2a8-config\") pod \"etcd-operator-b45778765-nh2kl\" (UID: \"d33f1ce9-b084-4ef3-887b-84f2d984f2a8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088245 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d33f1ce9-b084-4ef3-887b-84f2d984f2a8-etcd-ca\") pod \"etcd-operator-b45778765-nh2kl\" (UID: \"d33f1ce9-b084-4ef3-887b-84f2d984f2a8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088268 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/88c056b9-7ee2-40be-b5ec-abec9f736e3c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-29vlk\" (UID: \"88c056b9-7ee2-40be-b5ec-abec9f736e3c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088290 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98e0e18f-d88b-416a-b18b-eaaea13c8e21-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cr9nk\" (UID: \"98e0e18f-d88b-416a-b18b-eaaea13c8e21\") " pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088314 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8c5d984b-9b31-4997-904a-f27989f78a8c-registry-certificates\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088338 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fd37087b-7ebe-4f7f-9121-d55c058d804b-metrics-tls\") pod \"dns-operator-744455d44c-79vhf\" (UID: \"fd37087b-7ebe-4f7f-9121-d55c058d804b\") " pod="openshift-dns-operator/dns-operator-744455d44c-79vhf" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088379 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqqmg\" (UniqueName: \"kubernetes.io/projected/ac950124-b8dd-4603-a82e-6e2484ce2aaa-kube-api-access-rqqmg\") pod \"migrator-59844c95c7-kk8nf\" (UID: \"ac950124-b8dd-4603-a82e-6e2484ce2aaa\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kk8nf" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088523 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8c5d984b-9b31-4997-904a-f27989f78a8c-registry-tls\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088546 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7635f2d4-3be1-490a-99b9-3783a3fc58a4-node-bootstrap-token\") pod \"machine-config-server-4sl2t\" (UID: \"7635f2d4-3be1-490a-99b9-3783a3fc58a4\") " pod="openshift-machine-config-operator/machine-config-server-4sl2t" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088564 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d27bd203-adaf-4959-8907-416cd199df98-csi-data-dir\") pod \"csi-hostpathplugin-s684g\" (UID: \"d27bd203-adaf-4959-8907-416cd199df98\") " pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088602 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7583c399-65b2-4e40-91eb-1908753b6a29-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9lzl9\" (UID: \"7583c399-65b2-4e40-91eb-1908753b6a29\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9lzl9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088622 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjgbm\" (UniqueName: \"kubernetes.io/projected/d27bd203-adaf-4959-8907-416cd199df98-kube-api-access-rjgbm\") pod \"csi-hostpathplugin-s684g\" (UID: \"d27bd203-adaf-4959-8907-416cd199df98\") " pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088639 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgnl2\" (UniqueName: \"kubernetes.io/projected/8c5d984b-9b31-4997-904a-f27989f78a8c-kube-api-access-dgnl2\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088654 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/06d8bfc5-8156-4181-845b-dde35a4be1fe-cert\") pod \"ingress-canary-cfw4c\" (UID: \"06d8bfc5-8156-4181-845b-dde35a4be1fe\") " pod="openshift-ingress-canary/ingress-canary-cfw4c" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088672 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzb85\" (UniqueName: \"kubernetes.io/projected/f5d5515f-a284-4278-8c00-9222ee11b428-kube-api-access-hzb85\") pod \"multus-admission-controller-857f4d67dd-hk4bx\" (UID: \"f5d5515f-a284-4278-8c00-9222ee11b428\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hk4bx" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.088689 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c5d984b-9b31-4997-904a-f27989f78a8c-trusted-ca\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.089033 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8c5d984b-9b31-4997-904a-f27989f78a8c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.089978 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w29tv\" (UniqueName: \"kubernetes.io/projected/98e0e18f-d88b-416a-b18b-eaaea13c8e21-kube-api-access-w29tv\") pod \"marketplace-operator-79b997595-cr9nk\" (UID: \"98e0e18f-d88b-416a-b18b-eaaea13c8e21\") " pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090091 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8c5d984b-9b31-4997-904a-f27989f78a8c-registry-certificates\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090098 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l85qq\" (UniqueName: \"kubernetes.io/projected/06d8bfc5-8156-4181-845b-dde35a4be1fe-kube-api-access-l85qq\") pod \"ingress-canary-cfw4c\" (UID: \"06d8bfc5-8156-4181-845b-dde35a4be1fe\") " pod="openshift-ingress-canary/ingress-canary-cfw4c" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090183 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d33f1ce9-b084-4ef3-887b-84f2d984f2a8-etcd-service-ca\") pod \"etcd-operator-b45778765-nh2kl\" (UID: \"d33f1ce9-b084-4ef3-887b-84f2d984f2a8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090273 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/611d3c60-91d0-45f5-a3aa-dffcc427ef0b-config\") pod \"service-ca-operator-777779d784-6hwbq\" (UID: \"611d3c60-91d0-45f5-a3aa-dffcc427ef0b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6hwbq" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090284 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d33f1ce9-b084-4ef3-887b-84f2d984f2a8-config\") pod \"etcd-operator-b45778765-nh2kl\" (UID: \"d33f1ce9-b084-4ef3-887b-84f2d984f2a8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090392 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d33f1ce9-b084-4ef3-887b-84f2d984f2a8-etcd-ca\") pod \"etcd-operator-b45778765-nh2kl\" (UID: \"d33f1ce9-b084-4ef3-887b-84f2d984f2a8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090543 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090576 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fc06abdb-2b20-4fbd-8766-7470b97b3e27-bound-sa-token\") pod \"ingress-operator-5b745b69d9-k2xl9\" (UID: \"fc06abdb-2b20-4fbd-8766-7470b97b3e27\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090604 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8c5d984b-9b31-4997-904a-f27989f78a8c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090629 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e73052a1-e668-48c1-8734-5a3dc1f6b5f2-config-volume\") pod \"dns-default-v6gj4\" (UID: \"e73052a1-e668-48c1-8734-5a3dc1f6b5f2\") " pod="openshift-dns/dns-default-v6gj4" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090652 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b09c8b8b-8696-4a22-8b9a-4125ab33bc33-proxy-tls\") pod \"machine-config-controller-84d6567774-852rn\" (UID: \"b09c8b8b-8696-4a22-8b9a-4125ab33bc33\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-852rn" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090675 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/98e0e18f-d88b-416a-b18b-eaaea13c8e21-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cr9nk\" (UID: \"98e0e18f-d88b-416a-b18b-eaaea13c8e21\") " pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090704 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjnrf\" (UniqueName: \"kubernetes.io/projected/2d266390-de7a-4456-ae91-f44fe6a8b532-kube-api-access-mjnrf\") pod \"catalog-operator-68c6474976-rlnls\" (UID: \"2d266390-de7a-4456-ae91-f44fe6a8b532\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090724 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/392432cc-4ff4-4466-acac-7298b2a6197b-webhook-cert\") pod \"packageserver-d55dfcdfc-hb9lv\" (UID: \"392432cc-4ff4-4466-acac-7298b2a6197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090780 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/611d3c60-91d0-45f5-a3aa-dffcc427ef0b-serving-cert\") pod \"service-ca-operator-777779d784-6hwbq\" (UID: \"611d3c60-91d0-45f5-a3aa-dffcc427ef0b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6hwbq" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090804 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fc06abdb-2b20-4fbd-8766-7470b97b3e27-trusted-ca\") pod \"ingress-operator-5b745b69d9-k2xl9\" (UID: \"fc06abdb-2b20-4fbd-8766-7470b97b3e27\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090834 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khkxd\" (UniqueName: \"kubernetes.io/projected/7583c399-65b2-4e40-91eb-1908753b6a29-kube-api-access-khkxd\") pod \"control-plane-machine-set-operator-78cbb6b69f-9lzl9\" (UID: \"7583c399-65b2-4e40-91eb-1908753b6a29\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9lzl9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090854 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aad8389e-af3c-4783-bf7d-12945fc25070-serving-cert\") pod \"console-operator-58897d9998-nmzxv\" (UID: \"aad8389e-af3c-4783-bf7d-12945fc25070\") " pod="openshift-console-operator/console-operator-58897d9998-nmzxv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090898 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1f572606-6161-4ac4-a33a-8adaea8c071a-signing-cabundle\") pod \"service-ca-9c57cc56f-ppfnt\" (UID: \"1f572606-6161-4ac4-a33a-8adaea8c071a\") " pod="openshift-service-ca/service-ca-9c57cc56f-ppfnt" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090922 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a9c03eb-ba03-4b29-b3ad-585f9873973b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dzmm9\" (UID: \"8a9c03eb-ba03-4b29-b3ad-585f9873973b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmm9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.090964 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/030b390b-1901-4a54-98f8-bf241a955c22-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jzd75\" (UID: \"030b390b-1901-4a54-98f8-bf241a955c22\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jzd75" Nov 25 16:49:18 crc kubenswrapper[4802]: E1125 16:49:18.091016 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:18.591000668 +0000 UTC m=+141.735347854 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091140 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a9c03eb-ba03-4b29-b3ad-585f9873973b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dzmm9\" (UID: \"8a9c03eb-ba03-4b29-b3ad-585f9873973b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmm9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091165 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/030b390b-1901-4a54-98f8-bf241a955c22-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jzd75\" (UID: \"030b390b-1901-4a54-98f8-bf241a955c22\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jzd75" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091195 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qkx8\" (UniqueName: \"kubernetes.io/projected/fc06abdb-2b20-4fbd-8766-7470b97b3e27-kube-api-access-4qkx8\") pod \"ingress-operator-5b745b69d9-k2xl9\" (UID: \"fc06abdb-2b20-4fbd-8766-7470b97b3e27\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091242 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d27bd203-adaf-4959-8907-416cd199df98-plugins-dir\") pod \"csi-hostpathplugin-s684g\" (UID: \"d27bd203-adaf-4959-8907-416cd199df98\") " pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091270 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c5d984b-9b31-4997-904a-f27989f78a8c-bound-sa-token\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091310 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/88c056b9-7ee2-40be-b5ec-abec9f736e3c-srv-cert\") pod \"olm-operator-6b444d44fb-29vlk\" (UID: \"88c056b9-7ee2-40be-b5ec-abec9f736e3c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091541 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/392432cc-4ff4-4466-acac-7298b2a6197b-tmpfs\") pod \"packageserver-d55dfcdfc-hb9lv\" (UID: \"392432cc-4ff4-4466-acac-7298b2a6197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091583 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/030b390b-1901-4a54-98f8-bf241a955c22-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jzd75\" (UID: \"030b390b-1901-4a54-98f8-bf241a955c22\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jzd75" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091606 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2852a9e-0047-4ae1-941d-dde0a4cc991f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-zjrnl\" (UID: \"c2852a9e-0047-4ae1-941d-dde0a4cc991f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zjrnl" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091632 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9tt57\" (UID: \"042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9tt57" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091663 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d3ecea3d-c41f-46f4-bf0d-d43156542fa6-images\") pod \"machine-config-operator-74547568cd-c8xrm\" (UID: \"d3ecea3d-c41f-46f4-bf0d-d43156542fa6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091688 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/392432cc-4ff4-4466-acac-7298b2a6197b-apiservice-cert\") pod \"packageserver-d55dfcdfc-hb9lv\" (UID: \"392432cc-4ff4-4466-acac-7298b2a6197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091715 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e73052a1-e668-48c1-8734-5a3dc1f6b5f2-metrics-tls\") pod \"dns-default-v6gj4\" (UID: \"e73052a1-e668-48c1-8734-5a3dc1f6b5f2\") " pod="openshift-dns/dns-default-v6gj4" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091746 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d3ecea3d-c41f-46f4-bf0d-d43156542fa6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-c8xrm\" (UID: \"d3ecea3d-c41f-46f4-bf0d-d43156542fa6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091775 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9-config\") pod \"kube-apiserver-operator-766d6c64bb-9tt57\" (UID: \"042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9tt57" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091805 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1f572606-6161-4ac4-a33a-8adaea8c071a-signing-key\") pod \"service-ca-9c57cc56f-ppfnt\" (UID: \"1f572606-6161-4ac4-a33a-8adaea8c071a\") " pod="openshift-service-ca/service-ca-9c57cc56f-ppfnt" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091833 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9tt57\" (UID: \"042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9tt57" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091858 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aad8389e-af3c-4783-bf7d-12945fc25070-trusted-ca\") pod \"console-operator-58897d9998-nmzxv\" (UID: \"aad8389e-af3c-4783-bf7d-12945fc25070\") " pod="openshift-console-operator/console-operator-58897d9998-nmzxv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091917 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aad8389e-af3c-4783-bf7d-12945fc25070-config\") pod \"console-operator-58897d9998-nmzxv\" (UID: \"aad8389e-af3c-4783-bf7d-12945fc25070\") " pod="openshift-console-operator/console-operator-58897d9998-nmzxv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091992 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b09c8b8b-8696-4a22-8b9a-4125ab33bc33-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-852rn\" (UID: \"b09c8b8b-8696-4a22-8b9a-4125ab33bc33\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-852rn" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.092027 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdn88\" (UniqueName: \"kubernetes.io/projected/1f572606-6161-4ac4-a33a-8adaea8c071a-kube-api-access-wdn88\") pod \"service-ca-9c57cc56f-ppfnt\" (UID: \"1f572606-6161-4ac4-a33a-8adaea8c071a\") " pod="openshift-service-ca/service-ca-9c57cc56f-ppfnt" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.092059 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d27bd203-adaf-4959-8907-416cd199df98-socket-dir\") pod \"csi-hostpathplugin-s684g\" (UID: \"d27bd203-adaf-4959-8907-416cd199df98\") " pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.092098 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sgsx\" (UniqueName: \"kubernetes.io/projected/fd37087b-7ebe-4f7f-9121-d55c058d804b-kube-api-access-6sgsx\") pod \"dns-operator-744455d44c-79vhf\" (UID: \"fd37087b-7ebe-4f7f-9121-d55c058d804b\") " pod="openshift-dns-operator/dns-operator-744455d44c-79vhf" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.092166 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7635f2d4-3be1-490a-99b9-3783a3fc58a4-certs\") pod \"machine-config-server-4sl2t\" (UID: \"7635f2d4-3be1-490a-99b9-3783a3fc58a4\") " pod="openshift-machine-config-operator/machine-config-server-4sl2t" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.092222 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64qpx\" (UniqueName: \"kubernetes.io/projected/88c056b9-7ee2-40be-b5ec-abec9f736e3c-kube-api-access-64qpx\") pod \"olm-operator-6b444d44fb-29vlk\" (UID: \"88c056b9-7ee2-40be-b5ec-abec9f736e3c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.092251 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d33f1ce9-b084-4ef3-887b-84f2d984f2a8-etcd-client\") pod \"etcd-operator-b45778765-nh2kl\" (UID: \"d33f1ce9-b084-4ef3-887b-84f2d984f2a8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.092312 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d27bd203-adaf-4959-8907-416cd199df98-mountpoint-dir\") pod \"csi-hostpathplugin-s684g\" (UID: \"d27bd203-adaf-4959-8907-416cd199df98\") " pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.092361 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzp7w\" (UniqueName: \"kubernetes.io/projected/c2852a9e-0047-4ae1-941d-dde0a4cc991f-kube-api-access-nzp7w\") pod \"package-server-manager-789f6589d5-zjrnl\" (UID: \"c2852a9e-0047-4ae1-941d-dde0a4cc991f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zjrnl" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.092429 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wnkh\" (UniqueName: \"kubernetes.io/projected/8a9c03eb-ba03-4b29-b3ad-585f9873973b-kube-api-access-6wnkh\") pod \"kube-storage-version-migrator-operator-b67b599dd-dzmm9\" (UID: \"8a9c03eb-ba03-4b29-b3ad-585f9873973b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmm9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.092513 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2d266390-de7a-4456-ae91-f44fe6a8b532-srv-cert\") pod \"catalog-operator-68c6474976-rlnls\" (UID: \"2d266390-de7a-4456-ae91-f44fe6a8b532\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.092548 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mv96\" (UniqueName: \"kubernetes.io/projected/d33f1ce9-b084-4ef3-887b-84f2d984f2a8-kube-api-access-4mv96\") pod \"etcd-operator-b45778765-nh2kl\" (UID: \"d33f1ce9-b084-4ef3-887b-84f2d984f2a8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.092580 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d3ecea3d-c41f-46f4-bf0d-d43156542fa6-images\") pod \"machine-config-operator-74547568cd-c8xrm\" (UID: \"d3ecea3d-c41f-46f4-bf0d-d43156542fa6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.092594 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t22ct\" (UniqueName: \"kubernetes.io/projected/d3ecea3d-c41f-46f4-bf0d-d43156542fa6-kube-api-access-t22ct\") pod \"machine-config-operator-74547568cd-c8xrm\" (UID: \"d3ecea3d-c41f-46f4-bf0d-d43156542fa6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.092676 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt9qm\" (UniqueName: \"kubernetes.io/projected/611d3c60-91d0-45f5-a3aa-dffcc427ef0b-kube-api-access-kt9qm\") pod \"service-ca-operator-777779d784-6hwbq\" (UID: \"611d3c60-91d0-45f5-a3aa-dffcc427ef0b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6hwbq" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.092760 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st6fd\" (UniqueName: \"kubernetes.io/projected/aad8389e-af3c-4783-bf7d-12945fc25070-kube-api-access-st6fd\") pod \"console-operator-58897d9998-nmzxv\" (UID: \"aad8389e-af3c-4783-bf7d-12945fc25070\") " pod="openshift-console-operator/console-operator-58897d9998-nmzxv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.092794 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28bbp\" (UniqueName: \"kubernetes.io/projected/b09c8b8b-8696-4a22-8b9a-4125ab33bc33-kube-api-access-28bbp\") pod \"machine-config-controller-84d6567774-852rn\" (UID: \"b09c8b8b-8696-4a22-8b9a-4125ab33bc33\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-852rn" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.092817 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt8nb\" (UniqueName: \"kubernetes.io/projected/e73052a1-e668-48c1-8734-5a3dc1f6b5f2-kube-api-access-rt8nb\") pod \"dns-default-v6gj4\" (UID: \"e73052a1-e668-48c1-8734-5a3dc1f6b5f2\") " pod="openshift-dns/dns-default-v6gj4" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.091769 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e73052a1-e668-48c1-8734-5a3dc1f6b5f2-config-volume\") pod \"dns-default-v6gj4\" (UID: \"e73052a1-e668-48c1-8734-5a3dc1f6b5f2\") " pod="openshift-dns/dns-default-v6gj4" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.093915 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9-config\") pod \"kube-apiserver-operator-766d6c64bb-9tt57\" (UID: \"042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9tt57" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.094613 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d33f1ce9-b084-4ef3-887b-84f2d984f2a8-serving-cert\") pod \"etcd-operator-b45778765-nh2kl\" (UID: \"d33f1ce9-b084-4ef3-887b-84f2d984f2a8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.094819 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98e0e18f-d88b-416a-b18b-eaaea13c8e21-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cr9nk\" (UID: \"98e0e18f-d88b-416a-b18b-eaaea13c8e21\") " pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.095498 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/392432cc-4ff4-4466-acac-7298b2a6197b-tmpfs\") pod \"packageserver-d55dfcdfc-hb9lv\" (UID: \"392432cc-4ff4-4466-acac-7298b2a6197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.096638 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a9c03eb-ba03-4b29-b3ad-585f9873973b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dzmm9\" (UID: \"8a9c03eb-ba03-4b29-b3ad-585f9873973b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmm9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.097214 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1f572606-6161-4ac4-a33a-8adaea8c071a-signing-cabundle\") pod \"service-ca-9c57cc56f-ppfnt\" (UID: \"1f572606-6161-4ac4-a33a-8adaea8c071a\") " pod="openshift-service-ca/service-ca-9c57cc56f-ppfnt" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.097461 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fd37087b-7ebe-4f7f-9121-d55c058d804b-metrics-tls\") pod \"dns-operator-744455d44c-79vhf\" (UID: \"fd37087b-7ebe-4f7f-9121-d55c058d804b\") " pod="openshift-dns-operator/dns-operator-744455d44c-79vhf" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.097491 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7583c399-65b2-4e40-91eb-1908753b6a29-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9lzl9\" (UID: \"7583c399-65b2-4e40-91eb-1908753b6a29\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9lzl9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.097618 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d3ecea3d-c41f-46f4-bf0d-d43156542fa6-proxy-tls\") pod \"machine-config-operator-74547568cd-c8xrm\" (UID: \"d3ecea3d-c41f-46f4-bf0d-d43156542fa6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.097970 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/030b390b-1901-4a54-98f8-bf241a955c22-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jzd75\" (UID: \"030b390b-1901-4a54-98f8-bf241a955c22\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jzd75" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.098239 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/98e0e18f-d88b-416a-b18b-eaaea13c8e21-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cr9nk\" (UID: \"98e0e18f-d88b-416a-b18b-eaaea13c8e21\") " pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.098273 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fc06abdb-2b20-4fbd-8766-7470b97b3e27-trusted-ca\") pod \"ingress-operator-5b745b69d9-k2xl9\" (UID: \"fc06abdb-2b20-4fbd-8766-7470b97b3e27\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.098351 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b09c8b8b-8696-4a22-8b9a-4125ab33bc33-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-852rn\" (UID: \"b09c8b8b-8696-4a22-8b9a-4125ab33bc33\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-852rn" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.098556 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/06d8bfc5-8156-4181-845b-dde35a4be1fe-cert\") pod \"ingress-canary-cfw4c\" (UID: \"06d8bfc5-8156-4181-845b-dde35a4be1fe\") " pod="openshift-ingress-canary/ingress-canary-cfw4c" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.098989 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/88c056b9-7ee2-40be-b5ec-abec9f736e3c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-29vlk\" (UID: \"88c056b9-7ee2-40be-b5ec-abec9f736e3c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.100089 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d33f1ce9-b084-4ef3-887b-84f2d984f2a8-etcd-client\") pod \"etcd-operator-b45778765-nh2kl\" (UID: \"d33f1ce9-b084-4ef3-887b-84f2d984f2a8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.100205 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f5d5515f-a284-4278-8c00-9222ee11b428-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-hk4bx\" (UID: \"f5d5515f-a284-4278-8c00-9222ee11b428\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hk4bx" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.100146 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/611d3c60-91d0-45f5-a3aa-dffcc427ef0b-serving-cert\") pod \"service-ca-operator-777779d784-6hwbq\" (UID: \"611d3c60-91d0-45f5-a3aa-dffcc427ef0b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6hwbq" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.100350 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2d266390-de7a-4456-ae91-f44fe6a8b532-srv-cert\") pod \"catalog-operator-68c6474976-rlnls\" (UID: \"2d266390-de7a-4456-ae91-f44fe6a8b532\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.100400 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8c5d984b-9b31-4997-904a-f27989f78a8c-registry-tls\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.100434 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/392432cc-4ff4-4466-acac-7298b2a6197b-apiservice-cert\") pod \"packageserver-d55dfcdfc-hb9lv\" (UID: \"392432cc-4ff4-4466-acac-7298b2a6197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.100532 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b09c8b8b-8696-4a22-8b9a-4125ab33bc33-proxy-tls\") pod \"machine-config-controller-84d6567774-852rn\" (UID: \"b09c8b8b-8696-4a22-8b9a-4125ab33bc33\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-852rn" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.100655 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/392432cc-4ff4-4466-acac-7298b2a6197b-webhook-cert\") pod \"packageserver-d55dfcdfc-hb9lv\" (UID: \"392432cc-4ff4-4466-acac-7298b2a6197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.100863 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9tt57\" (UID: \"042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9tt57" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.101504 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2d266390-de7a-4456-ae91-f44fe6a8b532-profile-collector-cert\") pod \"catalog-operator-68c6474976-rlnls\" (UID: \"2d266390-de7a-4456-ae91-f44fe6a8b532\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.101612 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1f572606-6161-4ac4-a33a-8adaea8c071a-signing-key\") pod \"service-ca-9c57cc56f-ppfnt\" (UID: \"1f572606-6161-4ac4-a33a-8adaea8c071a\") " pod="openshift-service-ca/service-ca-9c57cc56f-ppfnt" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.101759 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aad8389e-af3c-4783-bf7d-12945fc25070-serving-cert\") pod \"console-operator-58897d9998-nmzxv\" (UID: \"aad8389e-af3c-4783-bf7d-12945fc25070\") " pod="openshift-console-operator/console-operator-58897d9998-nmzxv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.102391 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a9c03eb-ba03-4b29-b3ad-585f9873973b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dzmm9\" (UID: \"8a9c03eb-ba03-4b29-b3ad-585f9873973b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmm9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.102508 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fc06abdb-2b20-4fbd-8766-7470b97b3e27-metrics-tls\") pod \"ingress-operator-5b745b69d9-k2xl9\" (UID: \"fc06abdb-2b20-4fbd-8766-7470b97b3e27\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.103006 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/88c056b9-7ee2-40be-b5ec-abec9f736e3c-srv-cert\") pod \"olm-operator-6b444d44fb-29vlk\" (UID: \"88c056b9-7ee2-40be-b5ec-abec9f736e3c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.103096 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8c5d984b-9b31-4997-904a-f27989f78a8c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.107050 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/030b390b-1901-4a54-98f8-bf241a955c22-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jzd75\" (UID: \"030b390b-1901-4a54-98f8-bf241a955c22\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jzd75" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.108387 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6"] Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.130229 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnqsr\" (UniqueName: \"kubernetes.io/projected/392432cc-4ff4-4466-acac-7298b2a6197b-kube-api-access-nnqsr\") pod \"packageserver-d55dfcdfc-hb9lv\" (UID: \"392432cc-4ff4-4466-acac-7298b2a6197b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.150198 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqqmg\" (UniqueName: \"kubernetes.io/projected/ac950124-b8dd-4603-a82e-6e2484ce2aaa-kube-api-access-rqqmg\") pod \"migrator-59844c95c7-kk8nf\" (UID: \"ac950124-b8dd-4603-a82e-6e2484ce2aaa\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kk8nf" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.157718 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zpl2d" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.160526 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d33f1ce9-b084-4ef3-887b-84f2d984f2a8-etcd-service-ca\") pod \"etcd-operator-b45778765-nh2kl\" (UID: \"d33f1ce9-b084-4ef3-887b-84f2d984f2a8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.160962 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2852a9e-0047-4ae1-941d-dde0a4cc991f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-zjrnl\" (UID: \"c2852a9e-0047-4ae1-941d-dde0a4cc991f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zjrnl" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.164899 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aad8389e-af3c-4783-bf7d-12945fc25070-trusted-ca\") pod \"console-operator-58897d9998-nmzxv\" (UID: \"aad8389e-af3c-4783-bf7d-12945fc25070\") " pod="openshift-console-operator/console-operator-58897d9998-nmzxv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.166858 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bv6b6"] Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.167032 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e73052a1-e668-48c1-8734-5a3dc1f6b5f2-metrics-tls\") pod \"dns-default-v6gj4\" (UID: \"e73052a1-e668-48c1-8734-5a3dc1f6b5f2\") " pod="openshift-dns/dns-default-v6gj4" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.167634 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c5d984b-9b31-4997-904a-f27989f78a8c-trusted-ca\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.167942 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d3ecea3d-c41f-46f4-bf0d-d43156542fa6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-c8xrm\" (UID: \"d3ecea3d-c41f-46f4-bf0d-d43156542fa6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.175023 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sl8lb" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.176393 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgnl2\" (UniqueName: \"kubernetes.io/projected/8c5d984b-9b31-4997-904a-f27989f78a8c-kube-api-access-dgnl2\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.178494 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hzx2q"] Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.192176 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w29tv\" (UniqueName: \"kubernetes.io/projected/98e0e18f-d88b-416a-b18b-eaaea13c8e21-kube-api-access-w29tv\") pod \"marketplace-operator-79b997595-cr9nk\" (UID: \"98e0e18f-d88b-416a-b18b-eaaea13c8e21\") " pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.194048 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.194208 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7635f2d4-3be1-490a-99b9-3783a3fc58a4-node-bootstrap-token\") pod \"machine-config-server-4sl2t\" (UID: \"7635f2d4-3be1-490a-99b9-3783a3fc58a4\") " pod="openshift-machine-config-operator/machine-config-server-4sl2t" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.194238 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d27bd203-adaf-4959-8907-416cd199df98-csi-data-dir\") pod \"csi-hostpathplugin-s684g\" (UID: \"d27bd203-adaf-4959-8907-416cd199df98\") " pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.194285 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjgbm\" (UniqueName: \"kubernetes.io/projected/d27bd203-adaf-4959-8907-416cd199df98-kube-api-access-rjgbm\") pod \"csi-hostpathplugin-s684g\" (UID: \"d27bd203-adaf-4959-8907-416cd199df98\") " pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.194411 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d27bd203-adaf-4959-8907-416cd199df98-plugins-dir\") pod \"csi-hostpathplugin-s684g\" (UID: \"d27bd203-adaf-4959-8907-416cd199df98\") " pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.194464 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d27bd203-adaf-4959-8907-416cd199df98-socket-dir\") pod \"csi-hostpathplugin-s684g\" (UID: \"d27bd203-adaf-4959-8907-416cd199df98\") " pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.194485 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7635f2d4-3be1-490a-99b9-3783a3fc58a4-certs\") pod \"machine-config-server-4sl2t\" (UID: \"7635f2d4-3be1-490a-99b9-3783a3fc58a4\") " pod="openshift-machine-config-operator/machine-config-server-4sl2t" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.194509 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d27bd203-adaf-4959-8907-416cd199df98-mountpoint-dir\") pod \"csi-hostpathplugin-s684g\" (UID: \"d27bd203-adaf-4959-8907-416cd199df98\") " pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.194570 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d27bd203-adaf-4959-8907-416cd199df98-registration-dir\") pod \"csi-hostpathplugin-s684g\" (UID: \"d27bd203-adaf-4959-8907-416cd199df98\") " pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.194590 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52k5m\" (UniqueName: \"kubernetes.io/projected/7635f2d4-3be1-490a-99b9-3783a3fc58a4-kube-api-access-52k5m\") pod \"machine-config-server-4sl2t\" (UID: \"7635f2d4-3be1-490a-99b9-3783a3fc58a4\") " pod="openshift-machine-config-operator/machine-config-server-4sl2t" Nov 25 16:49:18 crc kubenswrapper[4802]: E1125 16:49:18.194796 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:18.694780269 +0000 UTC m=+141.839127455 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.196009 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d27bd203-adaf-4959-8907-416cd199df98-socket-dir\") pod \"csi-hostpathplugin-s684g\" (UID: \"d27bd203-adaf-4959-8907-416cd199df98\") " pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.196116 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d27bd203-adaf-4959-8907-416cd199df98-csi-data-dir\") pod \"csi-hostpathplugin-s684g\" (UID: \"d27bd203-adaf-4959-8907-416cd199df98\") " pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.196334 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d27bd203-adaf-4959-8907-416cd199df98-plugins-dir\") pod \"csi-hostpathplugin-s684g\" (UID: \"d27bd203-adaf-4959-8907-416cd199df98\") " pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.196385 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d27bd203-adaf-4959-8907-416cd199df98-mountpoint-dir\") pod \"csi-hostpathplugin-s684g\" (UID: \"d27bd203-adaf-4959-8907-416cd199df98\") " pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.196475 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d27bd203-adaf-4959-8907-416cd199df98-registration-dir\") pod \"csi-hostpathplugin-s684g\" (UID: \"d27bd203-adaf-4959-8907-416cd199df98\") " pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.198968 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7635f2d4-3be1-490a-99b9-3783a3fc58a4-node-bootstrap-token\") pod \"machine-config-server-4sl2t\" (UID: \"7635f2d4-3be1-490a-99b9-3783a3fc58a4\") " pod="openshift-machine-config-operator/machine-config-server-4sl2t" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.211098 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7635f2d4-3be1-490a-99b9-3783a3fc58a4-certs\") pod \"machine-config-server-4sl2t\" (UID: \"7635f2d4-3be1-490a-99b9-3783a3fc58a4\") " pod="openshift-machine-config-operator/machine-config-server-4sl2t" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.223246 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kk8nf" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.226580 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzb85\" (UniqueName: \"kubernetes.io/projected/f5d5515f-a284-4278-8c00-9222ee11b428-kube-api-access-hzb85\") pod \"multus-admission-controller-857f4d67dd-hk4bx\" (UID: \"f5d5515f-a284-4278-8c00-9222ee11b428\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hk4bx" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.234875 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fc06abdb-2b20-4fbd-8766-7470b97b3e27-bound-sa-token\") pod \"ingress-operator-5b745b69d9-k2xl9\" (UID: \"fc06abdb-2b20-4fbd-8766-7470b97b3e27\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.235681 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aad8389e-af3c-4783-bf7d-12945fc25070-config\") pod \"console-operator-58897d9998-nmzxv\" (UID: \"aad8389e-af3c-4783-bf7d-12945fc25070\") " pod="openshift-console-operator/console-operator-58897d9998-nmzxv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.256757 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2prtt"] Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.259226 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l85qq\" (UniqueName: \"kubernetes.io/projected/06d8bfc5-8156-4181-845b-dde35a4be1fe-kube-api-access-l85qq\") pod \"ingress-canary-cfw4c\" (UID: \"06d8bfc5-8156-4181-845b-dde35a4be1fe\") " pod="openshift-ingress-canary/ingress-canary-cfw4c" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.268249 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-smpjk"] Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.274608 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjnrf\" (UniqueName: \"kubernetes.io/projected/2d266390-de7a-4456-ae91-f44fe6a8b532-kube-api-access-mjnrf\") pod \"catalog-operator-68c6474976-rlnls\" (UID: \"2d266390-de7a-4456-ae91-f44fe6a8b532\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.297387 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.297656 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t22ct\" (UniqueName: \"kubernetes.io/projected/d3ecea3d-c41f-46f4-bf0d-d43156542fa6-kube-api-access-t22ct\") pod \"machine-config-operator-74547568cd-c8xrm\" (UID: \"d3ecea3d-c41f-46f4-bf0d-d43156542fa6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm" Nov 25 16:49:18 crc kubenswrapper[4802]: E1125 16:49:18.297797 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:18.797780329 +0000 UTC m=+141.942127585 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.297902 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.303258 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.317715 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-cfw4c" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.334289 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt8nb\" (UniqueName: \"kubernetes.io/projected/e73052a1-e668-48c1-8734-5a3dc1f6b5f2-kube-api-access-rt8nb\") pod \"dns-default-v6gj4\" (UID: \"e73052a1-e668-48c1-8734-5a3dc1f6b5f2\") " pod="openshift-dns/dns-default-v6gj4" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.352592 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64qpx\" (UniqueName: \"kubernetes.io/projected/88c056b9-7ee2-40be-b5ec-abec9f736e3c-kube-api-access-64qpx\") pod \"olm-operator-6b444d44fb-29vlk\" (UID: \"88c056b9-7ee2-40be-b5ec-abec9f736e3c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.359898 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c5d984b-9b31-4997-904a-f27989f78a8c-bound-sa-token\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.371253 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wnkh\" (UniqueName: \"kubernetes.io/projected/8a9c03eb-ba03-4b29-b3ad-585f9873973b-kube-api-access-6wnkh\") pod \"kube-storage-version-migrator-operator-b67b599dd-dzmm9\" (UID: \"8a9c03eb-ba03-4b29-b3ad-585f9873973b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmm9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.394563 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzp7w\" (UniqueName: \"kubernetes.io/projected/c2852a9e-0047-4ae1-941d-dde0a4cc991f-kube-api-access-nzp7w\") pod \"package-server-manager-789f6589d5-zjrnl\" (UID: \"c2852a9e-0047-4ae1-941d-dde0a4cc991f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zjrnl" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.398908 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:18 crc kubenswrapper[4802]: E1125 16:49:18.399489 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:18.899466913 +0000 UTC m=+142.043814099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.412923 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sgsx\" (UniqueName: \"kubernetes.io/projected/fd37087b-7ebe-4f7f-9121-d55c058d804b-kube-api-access-6sgsx\") pod \"dns-operator-744455d44c-79vhf\" (UID: \"fd37087b-7ebe-4f7f-9121-d55c058d804b\") " pod="openshift-dns-operator/dns-operator-744455d44c-79vhf" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.418164 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-79vhf" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.432383 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdn88\" (UniqueName: \"kubernetes.io/projected/1f572606-6161-4ac4-a33a-8adaea8c071a-kube-api-access-wdn88\") pod \"service-ca-9c57cc56f-ppfnt\" (UID: \"1f572606-6161-4ac4-a33a-8adaea8c071a\") " pod="openshift-service-ca/service-ca-9c57cc56f-ppfnt" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.453625 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khkxd\" (UniqueName: \"kubernetes.io/projected/7583c399-65b2-4e40-91eb-1908753b6a29-kube-api-access-khkxd\") pod \"control-plane-machine-set-operator-78cbb6b69f-9lzl9\" (UID: \"7583c399-65b2-4e40-91eb-1908753b6a29\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9lzl9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.474942 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st6fd\" (UniqueName: \"kubernetes.io/projected/aad8389e-af3c-4783-bf7d-12945fc25070-kube-api-access-st6fd\") pod \"console-operator-58897d9998-nmzxv\" (UID: \"aad8389e-af3c-4783-bf7d-12945fc25070\") " pod="openshift-console-operator/console-operator-58897d9998-nmzxv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.478935 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmm9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.486262 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-hk4bx" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.492237 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qkx8\" (UniqueName: \"kubernetes.io/projected/fc06abdb-2b20-4fbd-8766-7470b97b3e27-kube-api-access-4qkx8\") pod \"ingress-operator-5b745b69d9-k2xl9\" (UID: \"fc06abdb-2b20-4fbd-8766-7470b97b3e27\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.492477 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9lzl9" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.500895 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: E1125 16:49:18.501342 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:19.001329412 +0000 UTC m=+142.145676598 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.512791 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-nmzxv" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.513581 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt9qm\" (UniqueName: \"kubernetes.io/projected/611d3c60-91d0-45f5-a3aa-dffcc427ef0b-kube-api-access-kt9qm\") pod \"service-ca-operator-777779d784-6hwbq\" (UID: \"611d3c60-91d0-45f5-a3aa-dffcc427ef0b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6hwbq" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.530470 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.531951 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mv96\" (UniqueName: \"kubernetes.io/projected/d33f1ce9-b084-4ef3-887b-84f2d984f2a8-kube-api-access-4mv96\") pod \"etcd-operator-b45778765-nh2kl\" (UID: \"d33f1ce9-b084-4ef3-887b-84f2d984f2a8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.538842 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.551534 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28bbp\" (UniqueName: \"kubernetes.io/projected/b09c8b8b-8696-4a22-8b9a-4125ab33bc33-kube-api-access-28bbp\") pod \"machine-config-controller-84d6567774-852rn\" (UID: \"b09c8b8b-8696-4a22-8b9a-4125ab33bc33\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-852rn" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.556793 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zjrnl" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.562502 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6hwbq" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.568714 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-kk8nf"] Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.569003 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-ppfnt" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.580496 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9tt57\" (UID: \"042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9tt57" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.587794 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.603763 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:18 crc kubenswrapper[4802]: E1125 16:49:18.605036 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:19.10500766 +0000 UTC m=+142.249354846 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.607971 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh"] Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.609377 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sccr2"] Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.609887 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/030b390b-1901-4a54-98f8-bf241a955c22-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jzd75\" (UID: \"030b390b-1901-4a54-98f8-bf241a955c22\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jzd75" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.610397 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-v6gj4" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.621874 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sl8lb"] Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.621918 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zpl2d"] Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.658782 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-z8dgw"] Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.666418 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52k5m\" (UniqueName: \"kubernetes.io/projected/7635f2d4-3be1-490a-99b9-3783a3fc58a4-kube-api-access-52k5m\") pod \"machine-config-server-4sl2t\" (UID: \"7635f2d4-3be1-490a-99b9-3783a3fc58a4\") " pod="openshift-machine-config-operator/machine-config-server-4sl2t" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.666743 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjgbm\" (UniqueName: \"kubernetes.io/projected/d27bd203-adaf-4959-8907-416cd199df98-kube-api-access-rjgbm\") pod \"csi-hostpathplugin-s684g\" (UID: \"d27bd203-adaf-4959-8907-416cd199df98\") " pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.667765 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9qz4w"] Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.674138 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-p7zvz"] Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.674704 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mpcc"] Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.708870 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: E1125 16:49:18.709516 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:19.20949578 +0000 UTC m=+142.353842966 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.733226 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9" Nov 25 16:49:18 crc kubenswrapper[4802]: W1125 16:49:18.739187 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63972ba9_04aa_42e8_a91c_f27796ab40f4.slice/crio-4f99eed08d483d1ec4ec5f15763112b9d49f1c189d29a229550e517e23341c16 WatchSource:0}: Error finding container 4f99eed08d483d1ec4ec5f15763112b9d49f1c189d29a229550e517e23341c16: Status 404 returned error can't find the container with id 4f99eed08d483d1ec4ec5f15763112b9d49f1c189d29a229550e517e23341c16 Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.739426 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9tt57" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.746248 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.766042 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jzd75" Nov 25 16:49:18 crc kubenswrapper[4802]: W1125 16:49:18.787890 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf33868c_7e7c_4d67_b478_40ee1d7ef69d.slice/crio-fb655c200abcc6563374ea984c98423043f8d47a6aa3af1b80912e87a8467f4f WatchSource:0}: Error finding container fb655c200abcc6563374ea984c98423043f8d47a6aa3af1b80912e87a8467f4f: Status 404 returned error can't find the container with id fb655c200abcc6563374ea984c98423043f8d47a6aa3af1b80912e87a8467f4f Nov 25 16:49:18 crc kubenswrapper[4802]: W1125 16:49:18.789371 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91ee6068_d199_4b76_81fa_897155d5a38d.slice/crio-f581050cedac162c7b2b88ab01ce6c589c06194fcc13d4b10274f994e9ffb1ce WatchSource:0}: Error finding container f581050cedac162c7b2b88ab01ce6c589c06194fcc13d4b10274f994e9ffb1ce: Status 404 returned error can't find the container with id f581050cedac162c7b2b88ab01ce6c589c06194fcc13d4b10274f994e9ffb1ce Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.802353 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-852rn" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.812725 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:18 crc kubenswrapper[4802]: E1125 16:49:18.813214 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:19.313193509 +0000 UTC m=+142.457540695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.914417 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:18 crc kubenswrapper[4802]: E1125 16:49:18.914766 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:19.41475297 +0000 UTC m=+142.559100156 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.940701 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-s684g" Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.947201 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-cfw4c"] Nov 25 16:49:18 crc kubenswrapper[4802]: I1125 16:49:18.947446 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-4sl2t" Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.011831 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cr9nk"] Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.013326 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv"] Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.016429 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:19 crc kubenswrapper[4802]: E1125 16:49:19.016808 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:19.516788364 +0000 UTC m=+142.661135550 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.084594 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sl8lb" event={"ID":"13d7ba8a-bacc-4294-8e56-c3425f284c1a","Type":"ContainerStarted","Data":"359d458f2cb2c6ad1ebae57149e0a73da0da62d7919c60f039cbbea6b71b0a1a"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.088692 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mpcc" event={"ID":"91ee6068-d199-4b76-81fa-897155d5a38d","Type":"ContainerStarted","Data":"f581050cedac162c7b2b88ab01ce6c589c06194fcc13d4b10274f994e9ffb1ce"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.095720 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-79vhf"] Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.101162 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j2t9z" event={"ID":"ea30ecf0-8806-4c5a-9ff6-2738ca72e50b","Type":"ContainerStarted","Data":"2ab20e7dd8f9aa66dd1e2a2cacdd3770649a46eaeecd28f0dba3f129bbf16289"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.110362 4802 generic.go:334] "Generic (PLEG): container finished" podID="54c5e683-04bc-4241-8d13-17b415237707" containerID="a14f8dea36e3b547313fdf21c6e65082b137cf7a440a83840619812a230ee036" exitCode=0 Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.111448 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" event={"ID":"54c5e683-04bc-4241-8d13-17b415237707","Type":"ContainerDied","Data":"a14f8dea36e3b547313fdf21c6e65082b137cf7a440a83840619812a230ee036"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.113364 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-z8dgw" event={"ID":"8a3ebc4f-0491-4dc1-8c5d-adb851b6ef87","Type":"ContainerStarted","Data":"0f8a164c3f990702df9f5ffa7e295428aaee7010294b43264e7b5e41d3b7cd6c"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.115860 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" event={"ID":"b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0","Type":"ContainerStarted","Data":"b2c1bd91bd3943bb265e593582f82d5282d046afa9bc4bd8e01a318de44a2655"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.116436 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" event={"ID":"b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0","Type":"ContainerStarted","Data":"23766174dac35bd15d9c3591c2797efb7bfc8327e14b1de547e06d3e552b5260"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.117783 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:19 crc kubenswrapper[4802]: E1125 16:49:19.118140 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:19.618113348 +0000 UTC m=+142.762460534 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.124790 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-nr4zg" event={"ID":"f7d5d65a-b6dd-45dd-9f35-5e02a915a64f","Type":"ContainerStarted","Data":"275b02d4ce8e326159e265968453eebddc04e71dc880f87e4ead79a825df2418"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.126805 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kk8nf" event={"ID":"ac950124-b8dd-4603-a82e-6e2484ce2aaa","Type":"ContainerStarted","Data":"2aebc6252455e095da07aa6162561b47593483f98de476df8f19071b77a82cd6"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.128797 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zpl2d" event={"ID":"fbb37aa9-5a46-4ef3-9251-748a657e0933","Type":"ContainerStarted","Data":"c53cb61bd4836439720a9c116983a1914c2ed6649d0c74eee0acb713f2d2db6f"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.133985 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" event={"ID":"49937863-524b-4ee0-b3c4-6522995947c4","Type":"ContainerStarted","Data":"762effa9f3301816aee979f8a5a5131b381fbc426b98fa762ae9165bf218ba29"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.134045 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" event={"ID":"49937863-524b-4ee0-b3c4-6522995947c4","Type":"ContainerStarted","Data":"bfdb812bd911c9d957a5ccbd08f626783b75cb2136935ae1c398341f5a3bcb8c"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.134803 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.138324 4802 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-bv6b6 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.138401 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" podUID="49937863-524b-4ee0-b3c4-6522995947c4" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.170961 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" event={"ID":"63972ba9-04aa-42e8-a91c-f27796ab40f4","Type":"ContainerStarted","Data":"4f99eed08d483d1ec4ec5f15763112b9d49f1c189d29a229550e517e23341c16"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.172860 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6" event={"ID":"7a17670c-87d6-4fc7-b197-6817f4467377","Type":"ContainerStarted","Data":"ebc3f1f0304983ccef28560e722415df8be56dbcde157b4161f1debb9046de05"} Nov 25 16:49:19 crc kubenswrapper[4802]: W1125 16:49:19.173336 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod392432cc_4ff4_4466_acac_7298b2a6197b.slice/crio-c9e66ef5c856235d97d267f32d5c7d4e584caf4c2c4ce1d974772745998c5334 WatchSource:0}: Error finding container c9e66ef5c856235d97d267f32d5c7d4e584caf4c2c4ce1d974772745998c5334: Status 404 returned error can't find the container with id c9e66ef5c856235d97d267f32d5c7d4e584caf4c2c4ce1d974772745998c5334 Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.175252 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc" event={"ID":"c673008d-695e-44a5-8b6d-1d7c585b5b72","Type":"ContainerStarted","Data":"15b10766343c97eab50f8cb940865a346f4bddcc3dc6d3a532c32efac76e75db"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.182584 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-hk4bx"] Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.186678 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" event={"ID":"17422b05-1782-421b-ab25-4ce61d267fc2","Type":"ContainerStarted","Data":"43da95e700a9b1716cfbf18b8b617d01f07c43141ecb1458fe75d8650f2dd033"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.187765 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9lzl9"] Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.196161 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" event={"ID":"62710ae2-1843-4880-8444-b501a206e145","Type":"ContainerStarted","Data":"c939662372ff4a969007770a3f1bb0c979bc65ea2dd883427e78b622f6fb0dfe"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.207963 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" event={"ID":"a0a85d28-62a0-4ddc-ab49-c73b6db760a0","Type":"ContainerStarted","Data":"6edc3e654fa60a61381309e01edd8594519f65d9e6060abceb5f323bd04ff205"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.211351 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmm9"] Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.213094 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" event={"ID":"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40","Type":"ContainerStarted","Data":"9a9233ef4628da8f5364a8d2d6c7ed3ab76ff73e1831a51c12d5a76019822c31"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.213171 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.217572 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-p7zvz" event={"ID":"df33868c-7e7c-4d67-b478-40ee1d7ef69d","Type":"ContainerStarted","Data":"fb655c200abcc6563374ea984c98423043f8d47a6aa3af1b80912e87a8467f4f"} Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.218250 4802 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-zvb9w container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.218319 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" podUID="ef963c3e-4e73-4fe3-aed1-21d8b7d8be40" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.218564 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:19 crc kubenswrapper[4802]: E1125 16:49:19.219351 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:19.719296838 +0000 UTC m=+142.863644204 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:19 crc kubenswrapper[4802]: W1125 16:49:19.263936 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd37087b_7ebe_4f7f_9121_d55c058d804b.slice/crio-8825a28fdff2f22192d0b2525fc5ca513f96fe65482e69175f5cbf14858197be WatchSource:0}: Error finding container 8825a28fdff2f22192d0b2525fc5ca513f96fe65482e69175f5cbf14858197be: Status 404 returned error can't find the container with id 8825a28fdff2f22192d0b2525fc5ca513f96fe65482e69175f5cbf14858197be Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.266278 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zjrnl"] Nov 25 16:49:19 crc kubenswrapper[4802]: W1125 16:49:19.267876 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7583c399_65b2_4e40_91eb_1908753b6a29.slice/crio-02f1e1497bf60eb740b98f8808a36b002b27480b7fa0bbda3262d14a92bc3932 WatchSource:0}: Error finding container 02f1e1497bf60eb740b98f8808a36b002b27480b7fa0bbda3262d14a92bc3932: Status 404 returned error can't find the container with id 02f1e1497bf60eb740b98f8808a36b002b27480b7fa0bbda3262d14a92bc3932 Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.320625 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:19 crc kubenswrapper[4802]: E1125 16:49:19.322687 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:19.822661548 +0000 UTC m=+142.967008924 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.422703 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:19 crc kubenswrapper[4802]: E1125 16:49:19.422955 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:19.922922704 +0000 UTC m=+143.067269890 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.423059 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:19 crc kubenswrapper[4802]: E1125 16:49:19.424225 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:19.924199839 +0000 UTC m=+143.068547025 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.526400 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:19 crc kubenswrapper[4802]: E1125 16:49:19.526784 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:20.026767566 +0000 UTC m=+143.171114752 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.569627 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk"] Nov 25 16:49:19 crc kubenswrapper[4802]: W1125 16:49:19.595617 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7635f2d4_3be1_490a_99b9_3783a3fc58a4.slice/crio-1a7d9afbea983f43b786dce98ea5b51920c683feed27c1ca9dbe2a1ec888b8b5 WatchSource:0}: Error finding container 1a7d9afbea983f43b786dce98ea5b51920c683feed27c1ca9dbe2a1ec888b8b5: Status 404 returned error can't find the container with id 1a7d9afbea983f43b786dce98ea5b51920c683feed27c1ca9dbe2a1ec888b8b5 Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.596012 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-nh2kl"] Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.627709 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:19 crc kubenswrapper[4802]: E1125 16:49:19.628167 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:20.128146012 +0000 UTC m=+143.272493268 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:19 crc kubenswrapper[4802]: W1125 16:49:19.661443 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88c056b9_7ee2_40be_b5ec_abec9f736e3c.slice/crio-e9ce800744d37d75f79e59ed5eaf781ccbc4aaca3595a8b868f16fdc672369cd WatchSource:0}: Error finding container e9ce800744d37d75f79e59ed5eaf781ccbc4aaca3595a8b868f16fdc672369cd: Status 404 returned error can't find the container with id e9ce800744d37d75f79e59ed5eaf781ccbc4aaca3595a8b868f16fdc672369cd Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.728977 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:19 crc kubenswrapper[4802]: E1125 16:49:19.729273 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:20.229245381 +0000 UTC m=+143.373592567 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.738623 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:19 crc kubenswrapper[4802]: E1125 16:49:19.739495 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:20.239473548 +0000 UTC m=+143.383820734 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.812733 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.817973 4802 patch_prober.go:28] interesting pod/router-default-5444994796-nr4zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 16:49:19 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Nov 25 16:49:19 crc kubenswrapper[4802]: [+]process-running ok Nov 25 16:49:19 crc kubenswrapper[4802]: healthz check failed Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.818038 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nr4zg" podUID="f7d5d65a-b6dd-45dd-9f35-5e02a915a64f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.840138 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:19 crc kubenswrapper[4802]: E1125 16:49:19.840366 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:20.340330949 +0000 UTC m=+143.484678135 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.840725 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:19 crc kubenswrapper[4802]: E1125 16:49:19.841093 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:20.341081249 +0000 UTC m=+143.485428455 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.942288 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:19 crc kubenswrapper[4802]: E1125 16:49:19.943310 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:20.443292728 +0000 UTC m=+143.587639914 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.988168 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9"] Nov 25 16:49:19 crc kubenswrapper[4802]: I1125 16:49:19.992935 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jzd75"] Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.038252 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-nmzxv"] Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.047558 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm"] Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.048300 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:20 crc kubenswrapper[4802]: E1125 16:49:20.048699 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:20.548682953 +0000 UTC m=+143.693030139 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.056763 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-v6gj4"] Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.071537 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls"] Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.071625 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-s684g"] Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.080839 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9tt57"] Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.081456 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" podStartSLOduration=122.081431849 podStartE2EDuration="2m2.081431849s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:20.077818961 +0000 UTC m=+143.222166147" watchObservedRunningTime="2025-11-25 16:49:20.081431849 +0000 UTC m=+143.225779035" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.088492 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-852rn"] Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.101005 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6hwbq"] Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.118141 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" podStartSLOduration=123.118097712 podStartE2EDuration="2m3.118097712s" podCreationTimestamp="2025-11-25 16:47:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:20.116483319 +0000 UTC m=+143.260830505" watchObservedRunningTime="2025-11-25 16:49:20.118097712 +0000 UTC m=+143.262444898" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.147107 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" podStartSLOduration=122.147090078 podStartE2EDuration="2m2.147090078s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:20.146582025 +0000 UTC m=+143.290929211" watchObservedRunningTime="2025-11-25 16:49:20.147090078 +0000 UTC m=+143.291437264" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.149147 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:20 crc kubenswrapper[4802]: E1125 16:49:20.149627 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:20.649607086 +0000 UTC m=+143.793954272 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:20 crc kubenswrapper[4802]: W1125 16:49:20.170191 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode73052a1_e668_48c1_8734_5a3dc1f6b5f2.slice/crio-843fd278ce868160b1b81b4a481e6803a3292bfd19f819b5a0ca6945f3434c68 WatchSource:0}: Error finding container 843fd278ce868160b1b81b4a481e6803a3292bfd19f819b5a0ca6945f3434c68: Status 404 returned error can't find the container with id 843fd278ce868160b1b81b4a481e6803a3292bfd19f819b5a0ca6945f3434c68 Nov 25 16:49:20 crc kubenswrapper[4802]: W1125 16:49:20.178230 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod042b9d23_1b47_44e3_8ae1_f0bbea2ee4e9.slice/crio-817d050332f08e8e191710b24f6098111c281bda1b307a3c33e9ac729180380e WatchSource:0}: Error finding container 817d050332f08e8e191710b24f6098111c281bda1b307a3c33e9ac729180380e: Status 404 returned error can't find the container with id 817d050332f08e8e191710b24f6098111c281bda1b307a3c33e9ac729180380e Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.253247 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:20 crc kubenswrapper[4802]: E1125 16:49:20.253833 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:20.753816339 +0000 UTC m=+143.898163525 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.276712 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-nr4zg" podStartSLOduration=122.276690668 podStartE2EDuration="2m2.276690668s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:20.276211075 +0000 UTC m=+143.420558261" watchObservedRunningTime="2025-11-25 16:49:20.276690668 +0000 UTC m=+143.421037854" Nov 25 16:49:20 crc kubenswrapper[4802]: W1125 16:49:20.277816 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb09c8b8b_8696_4a22_8b9a_4125ab33bc33.slice/crio-c2ea3628d2fd8dfb9bde106899172dfef1006be5408650b4719edc98d5f34953 WatchSource:0}: Error finding container c2ea3628d2fd8dfb9bde106899172dfef1006be5408650b4719edc98d5f34953: Status 404 returned error can't find the container with id c2ea3628d2fd8dfb9bde106899172dfef1006be5408650b4719edc98d5f34953 Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.292729 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ppfnt"] Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.305669 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" event={"ID":"392432cc-4ff4-4466-acac-7298b2a6197b","Type":"ContainerStarted","Data":"c9e66ef5c856235d97d267f32d5c7d4e584caf4c2c4ce1d974772745998c5334"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.309888 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-79vhf" event={"ID":"fd37087b-7ebe-4f7f-9121-d55c058d804b","Type":"ContainerStarted","Data":"8825a28fdff2f22192d0b2525fc5ca513f96fe65482e69175f5cbf14858197be"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.318418 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-cfw4c" event={"ID":"06d8bfc5-8156-4181-845b-dde35a4be1fe","Type":"ContainerStarted","Data":"218e3dcf3306a7e7091fdbc0322cf00f03e172dcc71f551f4ce018c28ce2cc7a"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.318483 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-cfw4c" event={"ID":"06d8bfc5-8156-4181-845b-dde35a4be1fe","Type":"ContainerStarted","Data":"764d0693423a8c0061307ecf79ddb3584e21595204221b299af29106a8f2e1cd"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.325097 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9qz4w" event={"ID":"9af7072b-a939-4cdb-813f-1dccb589763a","Type":"ContainerStarted","Data":"d25bc65cc5ea5ad7175ec072f20a4a5adf263beeb8b627c780ffe30c412f3093"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.325179 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9qz4w" event={"ID":"9af7072b-a939-4cdb-813f-1dccb589763a","Type":"ContainerStarted","Data":"449a0965351e91963f248c792b4e3144be65cf2ca020b0b4451f5827f4d84dd6"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.329681 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" event={"ID":"d33f1ce9-b084-4ef3-887b-84f2d984f2a8","Type":"ContainerStarted","Data":"eda862428052a38f438fbba276213fda87a1f71956d4fb6cfb38ef8c80ac11a6"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.331870 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk" event={"ID":"88c056b9-7ee2-40be-b5ec-abec9f736e3c","Type":"ContainerStarted","Data":"e9ce800744d37d75f79e59ed5eaf781ccbc4aaca3595a8b868f16fdc672369cd"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.335000 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zpl2d" event={"ID":"fbb37aa9-5a46-4ef3-9251-748a657e0933","Type":"ContainerStarted","Data":"30a0464cbdcd5e34f6af502e5db922745f17e622c8bb03eb11cfa72156fa2fe0"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.345358 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mpcc" event={"ID":"91ee6068-d199-4b76-81fa-897155d5a38d","Type":"ContainerStarted","Data":"0c9da36789b97c8ef7decf817c6a8e9e0285f069914ede8f0f397031ebdb3970"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.354380 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:20 crc kubenswrapper[4802]: E1125 16:49:20.354885 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:20.854863905 +0000 UTC m=+143.999211081 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.355094 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6" podStartSLOduration=122.355066641 podStartE2EDuration="2m2.355066641s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:20.353723045 +0000 UTC m=+143.498070231" watchObservedRunningTime="2025-11-25 16:49:20.355066641 +0000 UTC m=+143.499413827" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.367433 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" event={"ID":"98e0e18f-d88b-416a-b18b-eaaea13c8e21","Type":"ContainerStarted","Data":"919559eccc834acc5ef1603d3125b827234982db0f9ad6449fa9be5dfd9d26ae"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.367904 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.367916 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" event={"ID":"98e0e18f-d88b-416a-b18b-eaaea13c8e21","Type":"ContainerStarted","Data":"e7c125bfb5eaf1161b88597f542893fcac84d0ce40086a4b12ad7c1977ed48cb"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.370399 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls" event={"ID":"2d266390-de7a-4456-ae91-f44fe6a8b532","Type":"ContainerStarted","Data":"593a78366f0c27249b8294c21baf2289931c071254206cae86dc53c6e7f7f51e"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.372851 4802 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cr9nk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.373261 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" podUID="98e0e18f-d88b-416a-b18b-eaaea13c8e21" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.374722 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-4sl2t" event={"ID":"7635f2d4-3be1-490a-99b9-3783a3fc58a4","Type":"ContainerStarted","Data":"1a7d9afbea983f43b786dce98ea5b51920c683feed27c1ca9dbe2a1ec888b8b5"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.388094 4802 generic.go:334] "Generic (PLEG): container finished" podID="b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0" containerID="b2c1bd91bd3943bb265e593582f82d5282d046afa9bc4bd8e01a318de44a2655" exitCode=0 Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.388300 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" event={"ID":"b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0","Type":"ContainerDied","Data":"b2c1bd91bd3943bb265e593582f82d5282d046afa9bc4bd8e01a318de44a2655"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.388762 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-cfw4c" podStartSLOduration=5.388742614 podStartE2EDuration="5.388742614s" podCreationTimestamp="2025-11-25 16:49:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:20.386498812 +0000 UTC m=+143.530846008" watchObservedRunningTime="2025-11-25 16:49:20.388742614 +0000 UTC m=+143.533089800" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.390008 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9" event={"ID":"fc06abdb-2b20-4fbd-8766-7470b97b3e27","Type":"ContainerStarted","Data":"1c0805c4d565143f0d5d23de05b810aea9e29ef6f047b261011b3ca0df6e2cf5"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.405402 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sl8lb" event={"ID":"13d7ba8a-bacc-4294-8e56-c3425f284c1a","Type":"ContainerStarted","Data":"20ade8320b4ad32c47da6ccbcc692c9755727ad8ce8df154acaf1aed65384b09"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.409330 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" event={"ID":"63972ba9-04aa-42e8-a91c-f27796ab40f4","Type":"ContainerStarted","Data":"bdad6a935ec9721095481ee741472175ca79bc15dff9508448f5f8185ea781be"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.410378 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.412532 4802 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-sccr2 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.33:6443/healthz\": dial tcp 10.217.0.33:6443: connect: connection refused" start-of-body= Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.412865 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" podUID="63972ba9-04aa-42e8-a91c-f27796ab40f4" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.33:6443/healthz\": dial tcp 10.217.0.33:6443: connect: connection refused" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.424538 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc" event={"ID":"c673008d-695e-44a5-8b6d-1d7c585b5b72","Type":"ContainerStarted","Data":"e58afaf9198bea8527ae88090042445cc7a59dcb853cbe3bd7e7e2ec01c32449"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.427423 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zpl2d" podStartSLOduration=122.42740157 podStartE2EDuration="2m2.42740157s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:20.425872339 +0000 UTC m=+143.570219545" watchObservedRunningTime="2025-11-25 16:49:20.42740157 +0000 UTC m=+143.571748756" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.434761 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9lzl9" event={"ID":"7583c399-65b2-4e40-91eb-1908753b6a29","Type":"ContainerStarted","Data":"02f1e1497bf60eb740b98f8808a36b002b27480b7fa0bbda3262d14a92bc3932"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.439591 4802 generic.go:334] "Generic (PLEG): container finished" podID="a0a85d28-62a0-4ddc-ab49-c73b6db760a0" containerID="90ebc3d6f9fa1f74126775a5cbc17bb6530c1176792e27fee1819671307b11b3" exitCode=0 Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.439688 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" event={"ID":"a0a85d28-62a0-4ddc-ab49-c73b6db760a0","Type":"ContainerDied","Data":"90ebc3d6f9fa1f74126775a5cbc17bb6530c1176792e27fee1819671307b11b3"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.439755 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" event={"ID":"a0a85d28-62a0-4ddc-ab49-c73b6db760a0","Type":"ContainerStarted","Data":"f6be850e321ecdd5299ba17c834d95109aabbc30fe1141a0eebc004b5133fc2f"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.443734 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2prtt" event={"ID":"62710ae2-1843-4880-8444-b501a206e145","Type":"ContainerStarted","Data":"14fae486af9ad6dad0055a12840fc3aec59bce41a53e2f21dbec14cb4bcf518d"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.447622 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" event={"ID":"17422b05-1782-421b-ab25-4ce61d267fc2","Type":"ContainerStarted","Data":"a686f2620d9bbe06f7f72afb13f5438c4d3ed7d2639d9a2f8dd8e80d01cb76a9"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.449784 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-v6gj4" event={"ID":"e73052a1-e668-48c1-8734-5a3dc1f6b5f2","Type":"ContainerStarted","Data":"843fd278ce868160b1b81b4a481e6803a3292bfd19f819b5a0ca6945f3434c68"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.450824 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9tt57" event={"ID":"042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9","Type":"ContainerStarted","Data":"817d050332f08e8e191710b24f6098111c281bda1b307a3c33e9ac729180380e"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.451883 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jzd75" event={"ID":"030b390b-1901-4a54-98f8-bf241a955c22","Type":"ContainerStarted","Data":"54ba71c583b483c9605100047f9e9621effbbf16211448a7c66487fcb3f6eecc"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.455530 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:20 crc kubenswrapper[4802]: E1125 16:49:20.456985 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:20.956964081 +0000 UTC m=+144.101311357 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.462782 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zjrnl" event={"ID":"c2852a9e-0047-4ae1-941d-dde0a4cc991f","Type":"ContainerStarted","Data":"13a53d9a258f4173fd07f068bac452af8aebedd60192ee85bfcbdf601c705b93"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.462903 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zjrnl" event={"ID":"c2852a9e-0047-4ae1-941d-dde0a4cc991f","Type":"ContainerStarted","Data":"42043e2af42cc706682973d15b0ae981c46adbede852b2bd0c9c0c9bfe11c25f"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.470353 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-z8dgw" event={"ID":"8a3ebc4f-0491-4dc1-8c5d-adb851b6ef87","Type":"ContainerStarted","Data":"f3e7dcb7ae343c32d482aa207c6509c79dcede730ffaa4f568a44524b3e135c6"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.471314 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mpcc" podStartSLOduration=123.471294959 podStartE2EDuration="2m3.471294959s" podCreationTimestamp="2025-11-25 16:47:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:20.470827906 +0000 UTC m=+143.615175092" watchObservedRunningTime="2025-11-25 16:49:20.471294959 +0000 UTC m=+143.615642145" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.476673 4802 patch_prober.go:28] interesting pod/downloads-7954f5f757-z8dgw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.476773 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-z8dgw" podUID="8a3ebc4f-0491-4dc1-8c5d-adb851b6ef87" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.484817 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-z8dgw" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.484879 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kk8nf" event={"ID":"ac950124-b8dd-4603-a82e-6e2484ce2aaa","Type":"ContainerStarted","Data":"1eda5fd03f9e1bfb3b1c84ec8f146bfe21e668030db868571e07eb715cabfe26"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.484908 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-p7zvz" event={"ID":"df33868c-7e7c-4d67-b478-40ee1d7ef69d","Type":"ContainerStarted","Data":"50e642d959a494e09f404cb7d8f56dd567585e7778fd7ae5b2cb76e698c92e93"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.484934 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmm9" event={"ID":"8a9c03eb-ba03-4b29-b3ad-585f9873973b","Type":"ContainerStarted","Data":"dab91111011bd5d3701aa6c89a1865d7696884efd7999ca8de367a4699f31a26"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.487160 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j2t9z" event={"ID":"ea30ecf0-8806-4c5a-9ff6-2738ca72e50b","Type":"ContainerStarted","Data":"322b3dfa575a911ce1879f9f30a247e03941a2311a8cb3e2093a6a682e59eab0"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.489374 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jkt6" event={"ID":"7a17670c-87d6-4fc7-b197-6817f4467377","Type":"ContainerStarted","Data":"21d6121bd92fe3f5b9b7ea16c67db6474ba17534465de04ea3364fe277bedab4"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.490708 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-hk4bx" event={"ID":"f5d5515f-a284-4278-8c00-9222ee11b428","Type":"ContainerStarted","Data":"835fcc749e67f56d608fc1e8ec07df0e8b7ce9b03fe0c11ef61cd39332ed2fa8"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.492078 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-nmzxv" event={"ID":"aad8389e-af3c-4783-bf7d-12945fc25070","Type":"ContainerStarted","Data":"223f2cbace4d5e4cbc99a9104bee47934898a4020ea724dfdae9535a372b0f85"} Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.493054 4802 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-bv6b6 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.493097 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" podUID="49937863-524b-4ee0-b3c4-6522995947c4" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.499640 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.510472 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sl8lb" podStartSLOduration=122.510452709 podStartE2EDuration="2m2.510452709s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:20.508856546 +0000 UTC m=+143.653203752" watchObservedRunningTime="2025-11-25 16:49:20.510452709 +0000 UTC m=+143.654799895" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.556921 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:20 crc kubenswrapper[4802]: E1125 16:49:20.557361 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:21.057285678 +0000 UTC m=+144.201632864 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.562380 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:20 crc kubenswrapper[4802]: E1125 16:49:20.565510 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:21.065486 +0000 UTC m=+144.209833376 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.625967 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" podStartSLOduration=122.625944768 podStartE2EDuration="2m2.625944768s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:20.623624445 +0000 UTC m=+143.767971651" watchObservedRunningTime="2025-11-25 16:49:20.625944768 +0000 UTC m=+143.770291954" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.670497 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:20 crc kubenswrapper[4802]: E1125 16:49:20.671432 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:21.171416759 +0000 UTC m=+144.315763945 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.772614 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:20 crc kubenswrapper[4802]: E1125 16:49:20.773012 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:21.272996661 +0000 UTC m=+144.417343847 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.783902 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2wcfc" podStartSLOduration=123.783883406 podStartE2EDuration="2m3.783883406s" podCreationTimestamp="2025-11-25 16:47:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:20.782016575 +0000 UTC m=+143.926363761" watchObservedRunningTime="2025-11-25 16:49:20.783883406 +0000 UTC m=+143.928230592" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.815553 4802 patch_prober.go:28] interesting pod/router-default-5444994796-nr4zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 16:49:20 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Nov 25 16:49:20 crc kubenswrapper[4802]: [+]process-running ok Nov 25 16:49:20 crc kubenswrapper[4802]: healthz check failed Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.815611 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nr4zg" podUID="f7d5d65a-b6dd-45dd-9f35-5e02a915a64f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.874730 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:20 crc kubenswrapper[4802]: E1125 16:49:20.875445 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:21.375418885 +0000 UTC m=+144.519766081 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.910068 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-z8dgw" podStartSLOduration=122.910028682 podStartE2EDuration="2m2.910028682s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:20.90735975 +0000 UTC m=+144.051706946" watchObservedRunningTime="2025-11-25 16:49:20.910028682 +0000 UTC m=+144.054375888" Nov 25 16:49:20 crc kubenswrapper[4802]: I1125 16:49:20.977431 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:20 crc kubenswrapper[4802]: E1125 16:49:20.978072 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:21.478054234 +0000 UTC m=+144.622401420 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.078532 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:21 crc kubenswrapper[4802]: E1125 16:49:21.079050 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:21.57901788 +0000 UTC m=+144.723365066 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.180385 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:21 crc kubenswrapper[4802]: E1125 16:49:21.180778 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:21.680763635 +0000 UTC m=+144.825110821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.282365 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:21 crc kubenswrapper[4802]: E1125 16:49:21.282504 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:21.78248243 +0000 UTC m=+144.926829616 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.283069 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:21 crc kubenswrapper[4802]: E1125 16:49:21.283432 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:21.783422646 +0000 UTC m=+144.927769832 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.285649 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" podStartSLOduration=123.285630886 podStartE2EDuration="2m3.285630886s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:21.285141433 +0000 UTC m=+144.429488609" watchObservedRunningTime="2025-11-25 16:49:21.285630886 +0000 UTC m=+144.429978072" Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.332847 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" podStartSLOduration=124.332824594 podStartE2EDuration="2m4.332824594s" podCreationTimestamp="2025-11-25 16:47:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:21.326613305 +0000 UTC m=+144.470960491" watchObservedRunningTime="2025-11-25 16:49:21.332824594 +0000 UTC m=+144.477171780" Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.383946 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:21 crc kubenswrapper[4802]: E1125 16:49:21.384352 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:21.884326019 +0000 UTC m=+145.028673225 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.384457 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:21 crc kubenswrapper[4802]: E1125 16:49:21.384869 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:21.884859603 +0000 UTC m=+145.029206799 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.486866 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:21 crc kubenswrapper[4802]: E1125 16:49:21.487709 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:21.987687568 +0000 UTC m=+145.132034744 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.529266 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-hk4bx" event={"ID":"f5d5515f-a284-4278-8c00-9222ee11b428","Type":"ContainerStarted","Data":"0ee546c841370e462980483d19d65c0a153751cbe60176eaec9795782df4301a"} Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.534002 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-4sl2t" event={"ID":"7635f2d4-3be1-490a-99b9-3783a3fc58a4","Type":"ContainerStarted","Data":"ce6e4b7d31c2afed86195446b41d2fbe4fd04045694af4da9182432347ea4494"} Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.543139 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-p7zvz" podStartSLOduration=123.543088129 podStartE2EDuration="2m3.543088129s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:21.54163008 +0000 UTC m=+144.685977276" watchObservedRunningTime="2025-11-25 16:49:21.543088129 +0000 UTC m=+144.687435315" Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.548467 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk" event={"ID":"88c056b9-7ee2-40be-b5ec-abec9f736e3c","Type":"ContainerStarted","Data":"1e2198551db501aa1c2b3ed72fcf9bbf753ea14ba2cc1436046b2375d93ad163"} Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.548833 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk" Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.552448 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-852rn" event={"ID":"b09c8b8b-8696-4a22-8b9a-4125ab33bc33","Type":"ContainerStarted","Data":"c2ea3628d2fd8dfb9bde106899172dfef1006be5408650b4719edc98d5f34953"} Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.559157 4802 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-29vlk container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.559226 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk" podUID="88c056b9-7ee2-40be-b5ec-abec9f736e3c" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.573543 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-j2t9z" podStartSLOduration=123.573522304 podStartE2EDuration="2m3.573522304s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:21.571349054 +0000 UTC m=+144.715696240" watchObservedRunningTime="2025-11-25 16:49:21.573522304 +0000 UTC m=+144.717869490" Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.589022 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:21 crc kubenswrapper[4802]: E1125 16:49:21.589451 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:22.089430114 +0000 UTC m=+145.233777370 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.591343 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9lzl9" event={"ID":"7583c399-65b2-4e40-91eb-1908753b6a29","Type":"ContainerStarted","Data":"6de591f72f50500bb8690e4936aee0bf2fb754b29b8fc181f9191303123f6353"} Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.597433 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s684g" event={"ID":"d27bd203-adaf-4959-8907-416cd199df98","Type":"ContainerStarted","Data":"1db4b998d9bc3dd51093f7485deb32b1c9efadd5b93c4813ff67bbcef3f0ca7b"} Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.602058 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" event={"ID":"392432cc-4ff4-4466-acac-7298b2a6197b","Type":"ContainerStarted","Data":"8f26f1a4440b048eaf3ada4d31b5f50a64f048db2733c50580f6b7a6f2345090"} Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.602410 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.610569 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6hwbq" event={"ID":"611d3c60-91d0-45f5-a3aa-dffcc427ef0b","Type":"ContainerStarted","Data":"5402a4461b8790286f0aaf01b01060cde77517d3671f137a1268284fc0bcdd65"} Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.613830 4802 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-hb9lv container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:5443/healthz\": dial tcp 10.217.0.42:5443: connect: connection refused" start-of-body= Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.613900 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" podUID="392432cc-4ff4-4466-acac-7298b2a6197b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.42:5443/healthz\": dial tcp 10.217.0.42:5443: connect: connection refused" Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.615549 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk" podStartSLOduration=123.615519791 podStartE2EDuration="2m3.615519791s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:21.60774809 +0000 UTC m=+144.752095276" watchObservedRunningTime="2025-11-25 16:49:21.615519791 +0000 UTC m=+144.759866977" Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.620980 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9qz4w" event={"ID":"9af7072b-a939-4cdb-813f-1dccb589763a","Type":"ContainerStarted","Data":"063b9edfeda873b73830b5262e463e3c7ad875c35498984a18cd9386529e85e9"} Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.624057 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-ppfnt" event={"ID":"1f572606-6161-4ac4-a33a-8adaea8c071a","Type":"ContainerStarted","Data":"ed9d08ee733006e9b65cb1b00f7d704b1d99a6374a6e2da125f998807ca7151a"} Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.635835 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" event={"ID":"54c5e683-04bc-4241-8d13-17b415237707","Type":"ContainerStarted","Data":"816f4316698c1647059133e5eb50fa153e32b25a24dacb77ad4e83689fa7ba20"} Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.643293 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm" event={"ID":"d3ecea3d-c41f-46f4-bf0d-d43156542fa6","Type":"ContainerStarted","Data":"e707f636f1573b3acb80c8a95381c3977cd59d5bbcfdcde95b22ecd6337bab09"} Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.652036 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9lzl9" podStartSLOduration=123.652016049 podStartE2EDuration="2m3.652016049s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:21.650626272 +0000 UTC m=+144.794973458" watchObservedRunningTime="2025-11-25 16:49:21.652016049 +0000 UTC m=+144.796363235" Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.664101 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmm9" event={"ID":"8a9c03eb-ba03-4b29-b3ad-585f9873973b","Type":"ContainerStarted","Data":"dc0982217b7c5eaede193ceb2a5409958ed35615ea96f0ccdd954580550cfc54"} Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.664876 4802 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-sccr2 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.33:6443/healthz\": dial tcp 10.217.0.33:6443: connect: connection refused" start-of-body= Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.664926 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" podUID="63972ba9-04aa-42e8-a91c-f27796ab40f4" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.33:6443/healthz\": dial tcp 10.217.0.33:6443: connect: connection refused" Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.665088 4802 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-bv6b6 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.665105 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" podUID="49937863-524b-4ee0-b3c4-6522995947c4" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.665188 4802 patch_prober.go:28] interesting pod/downloads-7954f5f757-z8dgw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.665210 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-z8dgw" podUID="8a3ebc4f-0491-4dc1-8c5d-adb851b6ef87" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.665435 4802 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cr9nk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.665454 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" podUID="98e0e18f-d88b-416a-b18b-eaaea13c8e21" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.689974 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:21 crc kubenswrapper[4802]: E1125 16:49:21.691750 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:22.191733285 +0000 UTC m=+145.336080471 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.695787 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" podStartSLOduration=123.695766254 podStartE2EDuration="2m3.695766254s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:21.692210427 +0000 UTC m=+144.836557623" watchObservedRunningTime="2025-11-25 16:49:21.695766254 +0000 UTC m=+144.840113430" Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.792102 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:21 crc kubenswrapper[4802]: E1125 16:49:21.792885 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:22.292867774 +0000 UTC m=+145.437214960 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.817792 4802 patch_prober.go:28] interesting pod/router-default-5444994796-nr4zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 16:49:21 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Nov 25 16:49:21 crc kubenswrapper[4802]: [+]process-running ok Nov 25 16:49:21 crc kubenswrapper[4802]: healthz check failed Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.817874 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nr4zg" podUID="f7d5d65a-b6dd-45dd-9f35-5e02a915a64f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 16:49:21 crc kubenswrapper[4802]: I1125 16:49:21.897906 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:21 crc kubenswrapper[4802]: E1125 16:49:21.898438 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:22.398416863 +0000 UTC m=+145.542764049 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:21.999512 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:22 crc kubenswrapper[4802]: E1125 16:49:21.999910 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:22.499892272 +0000 UTC m=+145.644239448 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.102701 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:22 crc kubenswrapper[4802]: E1125 16:49:22.121013 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:22.620976642 +0000 UTC m=+145.765323828 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.204807 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:22 crc kubenswrapper[4802]: E1125 16:49:22.205337 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:22.705317956 +0000 UTC m=+145.849665132 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.306818 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:22 crc kubenswrapper[4802]: E1125 16:49:22.307237 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:22.807220906 +0000 UTC m=+145.951568092 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.408510 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:22 crc kubenswrapper[4802]: E1125 16:49:22.408921 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:22.90890425 +0000 UTC m=+146.053251436 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.509416 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:22 crc kubenswrapper[4802]: E1125 16:49:22.510015 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:23.009993389 +0000 UTC m=+146.154340575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.611644 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:22 crc kubenswrapper[4802]: E1125 16:49:22.611917 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:23.111906288 +0000 UTC m=+146.256253474 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.674182 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm" event={"ID":"d3ecea3d-c41f-46f4-bf0d-d43156542fa6","Type":"ContainerStarted","Data":"5fb3db88469a3f73e137f4e91ecf1745583e146726f8b9bb60d9dddcc3619552"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.674222 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm" event={"ID":"d3ecea3d-c41f-46f4-bf0d-d43156542fa6","Type":"ContainerStarted","Data":"50e473b5bcf5da0f36e08aa4e747a48f53d6a5a0fafdd4d4964138d5789e465c"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.680318 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6hwbq" event={"ID":"611d3c60-91d0-45f5-a3aa-dffcc427ef0b","Type":"ContainerStarted","Data":"976a724e02d2f5477e0d065dccc87f0ad95b85fa5d7e3c0960ce9b0d707f78b0"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.686576 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-852rn" event={"ID":"b09c8b8b-8696-4a22-8b9a-4125ab33bc33","Type":"ContainerStarted","Data":"19229a22d3b8b4274cbd7a3896cd08071bc7fa4ba1cbcfb2b49be123f0928765"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.686612 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-852rn" event={"ID":"b09c8b8b-8696-4a22-8b9a-4125ab33bc33","Type":"ContainerStarted","Data":"c86b842a3a4cce6b1901d03cae559056f31696bbd76607852589ca1ca8f4ac7a"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.688853 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" event={"ID":"b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0","Type":"ContainerStarted","Data":"2fa2ae534038626b9255bf4845bd1c0b6754ee7ee3435a6fa52eef2eb63e23b2"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.689218 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.690439 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jzd75" event={"ID":"030b390b-1901-4a54-98f8-bf241a955c22","Type":"ContainerStarted","Data":"d9c0252af3ba3f3f324a85212e872a06149371b93ec29da63b8832c169da11d0"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.692682 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" event={"ID":"a0a85d28-62a0-4ddc-ab49-c73b6db760a0","Type":"ContainerStarted","Data":"41f42a4d81798ad63624f2ba49f340114fcb575c49fc6ad2523cc3cf1014c6f7"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.694713 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-hk4bx" event={"ID":"f5d5515f-a284-4278-8c00-9222ee11b428","Type":"ContainerStarted","Data":"8ccc0d37ade2367686719a73d113f17d2e5dd9355b1747143b902e4567ea75e2"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.696889 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-79vhf" event={"ID":"fd37087b-7ebe-4f7f-9121-d55c058d804b","Type":"ContainerStarted","Data":"c5ff10f2082683a1969f065620c5fffe8451bc4e3564ef9b8f17a6f445ca3292"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.696917 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-79vhf" event={"ID":"fd37087b-7ebe-4f7f-9121-d55c058d804b","Type":"ContainerStarted","Data":"0e8bc9cab20bef5cfcc6076a2cc22004f5bcdb9e0a1b3b1ce9e9f0cff672b5c2"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.697995 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c8xrm" podStartSLOduration=124.69797693 podStartE2EDuration="2m4.69797693s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:22.696369237 +0000 UTC m=+145.840716423" watchObservedRunningTime="2025-11-25 16:49:22.69797693 +0000 UTC m=+145.842324106" Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.698885 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zjrnl" event={"ID":"c2852a9e-0047-4ae1-941d-dde0a4cc991f","Type":"ContainerStarted","Data":"809d6989c3736e96d31d0557822749895ad8c64ea7d626b4f368855684fa494b"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.699513 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zjrnl" Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.700557 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9" event={"ID":"fc06abdb-2b20-4fbd-8766-7470b97b3e27","Type":"ContainerStarted","Data":"873c83f273a6745be521fc224f32e052c586f8e22f56e1eab27fd7bbdb02e114"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.700583 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9" event={"ID":"fc06abdb-2b20-4fbd-8766-7470b97b3e27","Type":"ContainerStarted","Data":"964d421bb887e9431e2a98f409655ff09c36c19a392a36a202822503c1c93803"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.703364 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls" event={"ID":"2d266390-de7a-4456-ae91-f44fe6a8b532","Type":"ContainerStarted","Data":"e0ca783b373cafaff8f2223a395ea1ae5c3dd46dad193616e790c303192042fa"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.703749 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls" Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.705056 4802 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-rlnls container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.705108 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls" podUID="2d266390-de7a-4456-ae91-f44fe6a8b532" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.705201 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" event={"ID":"d33f1ce9-b084-4ef3-887b-84f2d984f2a8","Type":"ContainerStarted","Data":"06abfd5d77469804bec80ec4455b6964c77485223ad5faf127e06793cdbea7d1"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.706451 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kk8nf" event={"ID":"ac950124-b8dd-4603-a82e-6e2484ce2aaa","Type":"ContainerStarted","Data":"620e835affbe34a704f3846fd6f0a3369fd1f2a48f11986fb5097a86ac81594d"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.707808 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-nmzxv" event={"ID":"aad8389e-af3c-4783-bf7d-12945fc25070","Type":"ContainerStarted","Data":"48c3b8ba8be07238a144227e62895ca25c15c643f6cac90a53ccb7bd9b83c07b"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.708506 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-nmzxv" Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.710008 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-v6gj4" event={"ID":"e73052a1-e668-48c1-8734-5a3dc1f6b5f2","Type":"ContainerStarted","Data":"310e2aa1c283185a8deb374e63d8e34bf2fdab593ece0999cb42162a9a8be258"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.710033 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-v6gj4" event={"ID":"e73052a1-e668-48c1-8734-5a3dc1f6b5f2","Type":"ContainerStarted","Data":"39443c84066021b9694dc453babefba9c3cbc8f2e2e5449c39f2c8614dfbfcba"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.710470 4802 patch_prober.go:28] interesting pod/console-operator-58897d9998-nmzxv container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/readyz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.710503 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-nmzxv" podUID="aad8389e-af3c-4783-bf7d-12945fc25070" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/readyz\": dial tcp 10.217.0.37:8443: connect: connection refused" Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.711430 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9tt57" event={"ID":"042b9d23-1b47-44e3-8ae1-f0bbea2ee4e9","Type":"ContainerStarted","Data":"2661a56dd63aebf930d603f5db83d3ca890f9572d1bbb95101136d1ffba96a27"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.712076 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:22 crc kubenswrapper[4802]: E1125 16:49:22.712262 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:23.212235636 +0000 UTC m=+146.356582822 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.712394 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:22 crc kubenswrapper[4802]: E1125 16:49:22.712766 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:23.21275588 +0000 UTC m=+146.357103076 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.714050 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-ppfnt" event={"ID":"1f572606-6161-4ac4-a33a-8adaea8c071a","Type":"ContainerStarted","Data":"b5f7be3e85d08eb23eb01b8fd725c1dd73dca9088b043f66d8c229438ee29794"} Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.717159 4802 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-hb9lv container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:5443/healthz\": dial tcp 10.217.0.42:5443: connect: connection refused" start-of-body= Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.717209 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" podUID="392432cc-4ff4-4466-acac-7298b2a6197b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.42:5443/healthz\": dial tcp 10.217.0.42:5443: connect: connection refused" Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.717479 4802 patch_prober.go:28] interesting pod/downloads-7954f5f757-z8dgw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.717499 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-z8dgw" podUID="8a3ebc4f-0491-4dc1-8c5d-adb851b6ef87" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.718387 4802 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-29vlk container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.718423 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk" podUID="88c056b9-7ee2-40be-b5ec-abec9f736e3c" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.734807 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-852rn" podStartSLOduration=124.734790717 podStartE2EDuration="2m4.734790717s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:22.732054423 +0000 UTC m=+145.876401599" watchObservedRunningTime="2025-11-25 16:49:22.734790717 +0000 UTC m=+145.879137903" Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.772175 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jzd75" podStartSLOduration=124.772159229 podStartE2EDuration="2m4.772159229s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:22.770396272 +0000 UTC m=+145.914743478" watchObservedRunningTime="2025-11-25 16:49:22.772159229 +0000 UTC m=+145.916506415" Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.797973 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" podStartSLOduration=125.797956638 podStartE2EDuration="2m5.797956638s" podCreationTimestamp="2025-11-25 16:47:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:22.795602895 +0000 UTC m=+145.939950081" watchObservedRunningTime="2025-11-25 16:49:22.797956638 +0000 UTC m=+145.942303824" Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.812936 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.816973 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" podStartSLOduration=125.816958553 podStartE2EDuration="2m5.816958553s" podCreationTimestamp="2025-11-25 16:47:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:22.81467256 +0000 UTC m=+145.959019756" watchObservedRunningTime="2025-11-25 16:49:22.816958553 +0000 UTC m=+145.961305739" Nov 25 16:49:22 crc kubenswrapper[4802]: E1125 16:49:22.818060 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:23.318047962 +0000 UTC m=+146.462395138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.827281 4802 patch_prober.go:28] interesting pod/router-default-5444994796-nr4zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 16:49:22 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Nov 25 16:49:22 crc kubenswrapper[4802]: [+]process-running ok Nov 25 16:49:22 crc kubenswrapper[4802]: healthz check failed Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.827624 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nr4zg" podUID="f7d5d65a-b6dd-45dd-9f35-5e02a915a64f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.869060 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6hwbq" podStartSLOduration=124.869037853 podStartE2EDuration="2m4.869037853s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:22.866863704 +0000 UTC m=+146.011210890" watchObservedRunningTime="2025-11-25 16:49:22.869037853 +0000 UTC m=+146.013385039" Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.887737 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-hk4bx" podStartSLOduration=124.887716979 podStartE2EDuration="2m4.887716979s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:22.886022463 +0000 UTC m=+146.030369649" watchObservedRunningTime="2025-11-25 16:49:22.887716979 +0000 UTC m=+146.032064175" Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.914913 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:22 crc kubenswrapper[4802]: E1125 16:49:22.915201 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:23.415187463 +0000 UTC m=+146.559534649 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.952594 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k2xl9" podStartSLOduration=124.952574276 podStartE2EDuration="2m4.952574276s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:22.912010507 +0000 UTC m=+146.056357693" watchObservedRunningTime="2025-11-25 16:49:22.952574276 +0000 UTC m=+146.096921462" Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.979670 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-ppfnt" podStartSLOduration=124.979653339 podStartE2EDuration="2m4.979653339s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:22.978767355 +0000 UTC m=+146.123114551" watchObservedRunningTime="2025-11-25 16:49:22.979653339 +0000 UTC m=+146.124000525" Nov 25 16:49:22 crc kubenswrapper[4802]: I1125 16:49:22.980146 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls" podStartSLOduration=124.980141522 podStartE2EDuration="2m4.980141522s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:22.951358673 +0000 UTC m=+146.095705859" watchObservedRunningTime="2025-11-25 16:49:22.980141522 +0000 UTC m=+146.124488708" Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.016331 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:23 crc kubenswrapper[4802]: E1125 16:49:23.016802 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:23.516785085 +0000 UTC m=+146.661132271 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.041661 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-nh2kl" podStartSLOduration=125.041646148 podStartE2EDuration="2m5.041646148s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:23.003619298 +0000 UTC m=+146.147966474" watchObservedRunningTime="2025-11-25 16:49:23.041646148 +0000 UTC m=+146.185993334" Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.042401 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zjrnl" podStartSLOduration=125.042395889 podStartE2EDuration="2m5.042395889s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:23.038512323 +0000 UTC m=+146.182859509" watchObservedRunningTime="2025-11-25 16:49:23.042395889 +0000 UTC m=+146.186743075" Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.097053 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kk8nf" podStartSLOduration=125.097034859 podStartE2EDuration="2m5.097034859s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:23.06201476 +0000 UTC m=+146.206361946" watchObservedRunningTime="2025-11-25 16:49:23.097034859 +0000 UTC m=+146.241382045" Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.114020 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-nmzxv" podStartSLOduration=125.113997028 podStartE2EDuration="2m5.113997028s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:23.112373204 +0000 UTC m=+146.256720390" watchObservedRunningTime="2025-11-25 16:49:23.113997028 +0000 UTC m=+146.258344214" Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.118935 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:23 crc kubenswrapper[4802]: E1125 16:49:23.119489 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:23.619472896 +0000 UTC m=+146.763820072 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.142670 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-4sl2t" podStartSLOduration=8.142634783 podStartE2EDuration="8.142634783s" podCreationTimestamp="2025-11-25 16:49:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:23.1391757 +0000 UTC m=+146.283522886" watchObservedRunningTime="2025-11-25 16:49:23.142634783 +0000 UTC m=+146.286981969" Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.181357 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9qz4w" podStartSLOduration=126.181337752 podStartE2EDuration="2m6.181337752s" podCreationTimestamp="2025-11-25 16:47:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:23.16464742 +0000 UTC m=+146.308994606" watchObservedRunningTime="2025-11-25 16:49:23.181337752 +0000 UTC m=+146.325684938" Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.220391 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:23 crc kubenswrapper[4802]: E1125 16:49:23.220749 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:23.720734009 +0000 UTC m=+146.865081195 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.223882 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dzmm9" podStartSLOduration=125.223869243 podStartE2EDuration="2m5.223869243s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:23.183196182 +0000 UTC m=+146.327543368" watchObservedRunningTime="2025-11-25 16:49:23.223869243 +0000 UTC m=+146.368216429" Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.224460 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9tt57" podStartSLOduration=125.2244543 podStartE2EDuration="2m5.2244543s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:23.222162077 +0000 UTC m=+146.366509253" watchObservedRunningTime="2025-11-25 16:49:23.2244543 +0000 UTC m=+146.368801486" Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.251027 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-79vhf" podStartSLOduration=125.251010009 podStartE2EDuration="2m5.251010009s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:23.250929627 +0000 UTC m=+146.395276813" watchObservedRunningTime="2025-11-25 16:49:23.251010009 +0000 UTC m=+146.395357195" Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.322175 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:23 crc kubenswrapper[4802]: E1125 16:49:23.322591 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:23.822577657 +0000 UTC m=+146.966924843 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.422963 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:23 crc kubenswrapper[4802]: E1125 16:49:23.423350 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:23.923332176 +0000 UTC m=+147.067679362 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.466254 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.525045 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:23 crc kubenswrapper[4802]: E1125 16:49:23.525489 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.025469273 +0000 UTC m=+147.169816459 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.581580 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" podStartSLOduration=125.581562512 podStartE2EDuration="2m5.581562512s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:23.299396789 +0000 UTC m=+146.443743985" watchObservedRunningTime="2025-11-25 16:49:23.581562512 +0000 UTC m=+146.725909698" Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.625772 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:23 crc kubenswrapper[4802]: E1125 16:49:23.626190 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.126174761 +0000 UTC m=+147.270521947 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.717631 4802 patch_prober.go:28] interesting pod/console-operator-58897d9998-nmzxv container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/readyz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.717688 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-nmzxv" podUID="aad8389e-af3c-4783-bf7d-12945fc25070" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/readyz\": dial tcp 10.217.0.37:8443: connect: connection refused" Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.717810 4802 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-rlnls container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.717862 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls" podUID="2d266390-de7a-4456-ae91-f44fe6a8b532" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.719441 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-v6gj4" Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.727533 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:23 crc kubenswrapper[4802]: E1125 16:49:23.727989 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.227963987 +0000 UTC m=+147.372311353 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.784227 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-v6gj4" podStartSLOduration=8.784190441 podStartE2EDuration="8.784190441s" podCreationTimestamp="2025-11-25 16:49:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:23.77900911 +0000 UTC m=+146.923356296" watchObservedRunningTime="2025-11-25 16:49:23.784190441 +0000 UTC m=+146.928537627" Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.816031 4802 patch_prober.go:28] interesting pod/router-default-5444994796-nr4zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 16:49:23 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Nov 25 16:49:23 crc kubenswrapper[4802]: [+]process-running ok Nov 25 16:49:23 crc kubenswrapper[4802]: healthz check failed Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.816089 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nr4zg" podUID="f7d5d65a-b6dd-45dd-9f35-5e02a915a64f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.830500 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:23 crc kubenswrapper[4802]: E1125 16:49:23.831971 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.331956024 +0000 UTC m=+147.476303210 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:23 crc kubenswrapper[4802]: I1125 16:49:23.932683 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:23 crc kubenswrapper[4802]: E1125 16:49:23.933037 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.433019632 +0000 UTC m=+147.577366818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.033471 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:24 crc kubenswrapper[4802]: E1125 16:49:24.033657 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.533615506 +0000 UTC m=+147.677962692 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.033761 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:24 crc kubenswrapper[4802]: E1125 16:49:24.034105 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.534091999 +0000 UTC m=+147.678439185 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.135254 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:24 crc kubenswrapper[4802]: E1125 16:49:24.135397 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.635376022 +0000 UTC m=+147.779723218 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.135461 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:24 crc kubenswrapper[4802]: E1125 16:49:24.135821 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.635811935 +0000 UTC m=+147.780159121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.235983 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:24 crc kubenswrapper[4802]: E1125 16:49:24.236189 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.736163342 +0000 UTC m=+147.880510528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.236558 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:24 crc kubenswrapper[4802]: E1125 16:49:24.237050 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.737028746 +0000 UTC m=+147.881375932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.249035 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.249146 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.337587 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:24 crc kubenswrapper[4802]: E1125 16:49:24.337824 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.837795105 +0000 UTC m=+147.982142291 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.337941 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:24 crc kubenswrapper[4802]: E1125 16:49:24.338263 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.838255327 +0000 UTC m=+147.982602513 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.438816 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:24 crc kubenswrapper[4802]: E1125 16:49:24.439087 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.939039637 +0000 UTC m=+148.083386833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.439220 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:24 crc kubenswrapper[4802]: E1125 16:49:24.439629 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:24.939621523 +0000 UTC m=+148.083968709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.541099 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:24 crc kubenswrapper[4802]: E1125 16:49:24.541307 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:25.041274456 +0000 UTC m=+148.185621642 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.541396 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.541440 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.541467 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.541510 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:49:24 crc kubenswrapper[4802]: E1125 16:49:24.541874 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:25.041830372 +0000 UTC m=+148.186177558 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.542371 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.548523 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.563258 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.616188 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.642066 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:24 crc kubenswrapper[4802]: E1125 16:49:24.642283 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:25.142240501 +0000 UTC m=+148.286587697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.642348 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.642584 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:49:24 crc kubenswrapper[4802]: E1125 16:49:24.642642 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:25.142629522 +0000 UTC m=+148.286976708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.652590 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.721996 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.735405 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.748769 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:24 crc kubenswrapper[4802]: E1125 16:49:24.749521 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:25.249494696 +0000 UTC m=+148.393841892 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.805564 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s684g" event={"ID":"d27bd203-adaf-4959-8907-416cd199df98","Type":"ContainerStarted","Data":"204e0fcfc5518f814646ddcc606b9e3910acbc3f9e4c0c63b15820b82464b7fc"} Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.807518 4802 patch_prober.go:28] interesting pod/console-operator-58897d9998-nmzxv container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/readyz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.807559 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-nmzxv" podUID="aad8389e-af3c-4783-bf7d-12945fc25070" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/readyz\": dial tcp 10.217.0.37:8443: connect: connection refused" Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.820323 4802 patch_prober.go:28] interesting pod/router-default-5444994796-nr4zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 16:49:24 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Nov 25 16:49:24 crc kubenswrapper[4802]: [+]process-running ok Nov 25 16:49:24 crc kubenswrapper[4802]: healthz check failed Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.820374 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nr4zg" podUID="f7d5d65a-b6dd-45dd-9f35-5e02a915a64f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.828521 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rlnls" Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.862867 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:24 crc kubenswrapper[4802]: E1125 16:49:24.863500 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:25.363484003 +0000 UTC m=+148.507831189 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:24 crc kubenswrapper[4802]: I1125 16:49:24.966972 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:24 crc kubenswrapper[4802]: E1125 16:49:24.968684 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:25.468663543 +0000 UTC m=+148.613010729 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.069985 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:25 crc kubenswrapper[4802]: E1125 16:49:25.070466 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:25.570449179 +0000 UTC m=+148.714796365 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.170590 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:25 crc kubenswrapper[4802]: E1125 16:49:25.171215 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:25.671199978 +0000 UTC m=+148.815547164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.271944 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:25 crc kubenswrapper[4802]: E1125 16:49:25.272303 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:25.772290256 +0000 UTC m=+148.916637442 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.373482 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:25 crc kubenswrapper[4802]: E1125 16:49:25.373958 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:25.873940599 +0000 UTC m=+149.018287785 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.475872 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:25 crc kubenswrapper[4802]: E1125 16:49:25.476237 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:25.976224 +0000 UTC m=+149.120571186 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.495641 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.496528 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.499501 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.499993 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.552228 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.576891 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:25 crc kubenswrapper[4802]: E1125 16:49:25.577248 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:26.077230795 +0000 UTC m=+149.221577981 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.583180 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r9nw2"] Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.584066 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r9nw2" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.588649 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.611391 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r9nw2"] Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.678241 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb067ba0-1c60-4b52-aa08-3675dd787fdf-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"cb067ba0-1c60-4b52-aa08-3675dd787fdf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.678406 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b19f412-72dd-4fe9-b0f1-3abc306653ae-catalog-content\") pod \"community-operators-r9nw2\" (UID: \"9b19f412-72dd-4fe9-b0f1-3abc306653ae\") " pod="openshift-marketplace/community-operators-r9nw2" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.678481 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm7mf\" (UniqueName: \"kubernetes.io/projected/9b19f412-72dd-4fe9-b0f1-3abc306653ae-kube-api-access-qm7mf\") pod \"community-operators-r9nw2\" (UID: \"9b19f412-72dd-4fe9-b0f1-3abc306653ae\") " pod="openshift-marketplace/community-operators-r9nw2" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.678526 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:25 crc kubenswrapper[4802]: E1125 16:49:25.678909 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:26.178896289 +0000 UTC m=+149.323243475 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.678966 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b19f412-72dd-4fe9-b0f1-3abc306653ae-utilities\") pod \"community-operators-r9nw2\" (UID: \"9b19f412-72dd-4fe9-b0f1-3abc306653ae\") " pod="openshift-marketplace/community-operators-r9nw2" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.678997 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cb067ba0-1c60-4b52-aa08-3675dd787fdf-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"cb067ba0-1c60-4b52-aa08-3675dd787fdf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.785680 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.785814 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb067ba0-1c60-4b52-aa08-3675dd787fdf-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"cb067ba0-1c60-4b52-aa08-3675dd787fdf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.785850 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b19f412-72dd-4fe9-b0f1-3abc306653ae-catalog-content\") pod \"community-operators-r9nw2\" (UID: \"9b19f412-72dd-4fe9-b0f1-3abc306653ae\") " pod="openshift-marketplace/community-operators-r9nw2" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.785876 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm7mf\" (UniqueName: \"kubernetes.io/projected/9b19f412-72dd-4fe9-b0f1-3abc306653ae-kube-api-access-qm7mf\") pod \"community-operators-r9nw2\" (UID: \"9b19f412-72dd-4fe9-b0f1-3abc306653ae\") " pod="openshift-marketplace/community-operators-r9nw2" Nov 25 16:49:25 crc kubenswrapper[4802]: E1125 16:49:25.785927 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:26.285900997 +0000 UTC m=+149.430248183 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.786029 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b19f412-72dd-4fe9-b0f1-3abc306653ae-utilities\") pod \"community-operators-r9nw2\" (UID: \"9b19f412-72dd-4fe9-b0f1-3abc306653ae\") " pod="openshift-marketplace/community-operators-r9nw2" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.786098 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cb067ba0-1c60-4b52-aa08-3675dd787fdf-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"cb067ba0-1c60-4b52-aa08-3675dd787fdf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.786280 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cb067ba0-1c60-4b52-aa08-3675dd787fdf-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"cb067ba0-1c60-4b52-aa08-3675dd787fdf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.786766 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b19f412-72dd-4fe9-b0f1-3abc306653ae-utilities\") pod \"community-operators-r9nw2\" (UID: \"9b19f412-72dd-4fe9-b0f1-3abc306653ae\") " pod="openshift-marketplace/community-operators-r9nw2" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.786783 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b19f412-72dd-4fe9-b0f1-3abc306653ae-catalog-content\") pod \"community-operators-r9nw2\" (UID: \"9b19f412-72dd-4fe9-b0f1-3abc306653ae\") " pod="openshift-marketplace/community-operators-r9nw2" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.806283 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rtgv8"] Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.807216 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rtgv8" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.821661 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.828014 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm7mf\" (UniqueName: \"kubernetes.io/projected/9b19f412-72dd-4fe9-b0f1-3abc306653ae-kube-api-access-qm7mf\") pod \"community-operators-r9nw2\" (UID: \"9b19f412-72dd-4fe9-b0f1-3abc306653ae\") " pod="openshift-marketplace/community-operators-r9nw2" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.828789 4802 patch_prober.go:28] interesting pod/router-default-5444994796-nr4zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 16:49:25 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Nov 25 16:49:25 crc kubenswrapper[4802]: [+]process-running ok Nov 25 16:49:25 crc kubenswrapper[4802]: healthz check failed Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.828861 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nr4zg" podUID="f7d5d65a-b6dd-45dd-9f35-5e02a915a64f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.829840 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"fd86ea90cd8a8227039000e0dd53bcc57dab4f824922267483ff004a85722922"} Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.832633 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"5b7c4ffc0df6e35f8c0246c504cb4fb3027ab35014b381628424ce4093c301d9"} Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.834150 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"97a470a63277f4273b10eb038baff3aba8691520d59945938c31a104e05cdb29"} Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.847654 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb067ba0-1c60-4b52-aa08-3675dd787fdf-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"cb067ba0-1c60-4b52-aa08-3675dd787fdf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.868144 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rtgv8"] Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.894890 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:25 crc kubenswrapper[4802]: E1125 16:49:25.895297 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:26.39528168 +0000 UTC m=+149.539628866 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.931193 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r9nw2" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.994242 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jrq5g"] Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.996348 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.996556 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/565aa173-3fac-4f4e-8fae-3fbe88184efb-catalog-content\") pod \"certified-operators-rtgv8\" (UID: \"565aa173-3fac-4f4e-8fae-3fbe88184efb\") " pod="openshift-marketplace/certified-operators-rtgv8" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.996584 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/565aa173-3fac-4f4e-8fae-3fbe88184efb-utilities\") pod \"certified-operators-rtgv8\" (UID: \"565aa173-3fac-4f4e-8fae-3fbe88184efb\") " pod="openshift-marketplace/certified-operators-rtgv8" Nov 25 16:49:25 crc kubenswrapper[4802]: I1125 16:49:25.996626 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h6dr\" (UniqueName: \"kubernetes.io/projected/565aa173-3fac-4f4e-8fae-3fbe88184efb-kube-api-access-4h6dr\") pod \"certified-operators-rtgv8\" (UID: \"565aa173-3fac-4f4e-8fae-3fbe88184efb\") " pod="openshift-marketplace/certified-operators-rtgv8" Nov 25 16:49:25 crc kubenswrapper[4802]: E1125 16:49:25.996726 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:26.496709708 +0000 UTC m=+149.641056884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.009280 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jrq5g" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.077541 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jrq5g"] Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.103990 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac5f001d-89fc-41e3-8a80-527e81eb55d0-catalog-content\") pod \"community-operators-jrq5g\" (UID: \"ac5f001d-89fc-41e3-8a80-527e81eb55d0\") " pod="openshift-marketplace/community-operators-jrq5g" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.104084 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac5f001d-89fc-41e3-8a80-527e81eb55d0-utilities\") pod \"community-operators-jrq5g\" (UID: \"ac5f001d-89fc-41e3-8a80-527e81eb55d0\") " pod="openshift-marketplace/community-operators-jrq5g" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.104156 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/565aa173-3fac-4f4e-8fae-3fbe88184efb-catalog-content\") pod \"certified-operators-rtgv8\" (UID: \"565aa173-3fac-4f4e-8fae-3fbe88184efb\") " pod="openshift-marketplace/certified-operators-rtgv8" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.104179 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/565aa173-3fac-4f4e-8fae-3fbe88184efb-utilities\") pod \"certified-operators-rtgv8\" (UID: \"565aa173-3fac-4f4e-8fae-3fbe88184efb\") " pod="openshift-marketplace/certified-operators-rtgv8" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.104235 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h6dr\" (UniqueName: \"kubernetes.io/projected/565aa173-3fac-4f4e-8fae-3fbe88184efb-kube-api-access-4h6dr\") pod \"certified-operators-rtgv8\" (UID: \"565aa173-3fac-4f4e-8fae-3fbe88184efb\") " pod="openshift-marketplace/certified-operators-rtgv8" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.104267 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxjt5\" (UniqueName: \"kubernetes.io/projected/ac5f001d-89fc-41e3-8a80-527e81eb55d0-kube-api-access-rxjt5\") pod \"community-operators-jrq5g\" (UID: \"ac5f001d-89fc-41e3-8a80-527e81eb55d0\") " pod="openshift-marketplace/community-operators-jrq5g" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.104332 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:26 crc kubenswrapper[4802]: E1125 16:49:26.104755 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:26.604741464 +0000 UTC m=+149.749088650 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.105331 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/565aa173-3fac-4f4e-8fae-3fbe88184efb-catalog-content\") pod \"certified-operators-rtgv8\" (UID: \"565aa173-3fac-4f4e-8fae-3fbe88184efb\") " pod="openshift-marketplace/certified-operators-rtgv8" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.105583 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/565aa173-3fac-4f4e-8fae-3fbe88184efb-utilities\") pod \"certified-operators-rtgv8\" (UID: \"565aa173-3fac-4f4e-8fae-3fbe88184efb\") " pod="openshift-marketplace/certified-operators-rtgv8" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.160930 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.173362 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nhwgv"] Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.178973 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nhwgv" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.181894 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h6dr\" (UniqueName: \"kubernetes.io/projected/565aa173-3fac-4f4e-8fae-3fbe88184efb-kube-api-access-4h6dr\") pod \"certified-operators-rtgv8\" (UID: \"565aa173-3fac-4f4e-8fae-3fbe88184efb\") " pod="openshift-marketplace/certified-operators-rtgv8" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.201078 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nhwgv"] Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.208182 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.208405 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac5f001d-89fc-41e3-8a80-527e81eb55d0-utilities\") pod \"community-operators-jrq5g\" (UID: \"ac5f001d-89fc-41e3-8a80-527e81eb55d0\") " pod="openshift-marketplace/community-operators-jrq5g" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.208458 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxjt5\" (UniqueName: \"kubernetes.io/projected/ac5f001d-89fc-41e3-8a80-527e81eb55d0-kube-api-access-rxjt5\") pod \"community-operators-jrq5g\" (UID: \"ac5f001d-89fc-41e3-8a80-527e81eb55d0\") " pod="openshift-marketplace/community-operators-jrq5g" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.208532 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac5f001d-89fc-41e3-8a80-527e81eb55d0-catalog-content\") pod \"community-operators-jrq5g\" (UID: \"ac5f001d-89fc-41e3-8a80-527e81eb55d0\") " pod="openshift-marketplace/community-operators-jrq5g" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.208955 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac5f001d-89fc-41e3-8a80-527e81eb55d0-catalog-content\") pod \"community-operators-jrq5g\" (UID: \"ac5f001d-89fc-41e3-8a80-527e81eb55d0\") " pod="openshift-marketplace/community-operators-jrq5g" Nov 25 16:49:26 crc kubenswrapper[4802]: E1125 16:49:26.209027 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:26.709007977 +0000 UTC m=+149.853355163 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.209249 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac5f001d-89fc-41e3-8a80-527e81eb55d0-utilities\") pod \"community-operators-jrq5g\" (UID: \"ac5f001d-89fc-41e3-8a80-527e81eb55d0\") " pod="openshift-marketplace/community-operators-jrq5g" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.282928 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxjt5\" (UniqueName: \"kubernetes.io/projected/ac5f001d-89fc-41e3-8a80-527e81eb55d0-kube-api-access-rxjt5\") pod \"community-operators-jrq5g\" (UID: \"ac5f001d-89fc-41e3-8a80-527e81eb55d0\") " pod="openshift-marketplace/community-operators-jrq5g" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.312453 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff-utilities\") pod \"certified-operators-nhwgv\" (UID: \"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff\") " pod="openshift-marketplace/certified-operators-nhwgv" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.312509 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dvqh\" (UniqueName: \"kubernetes.io/projected/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff-kube-api-access-8dvqh\") pod \"certified-operators-nhwgv\" (UID: \"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff\") " pod="openshift-marketplace/certified-operators-nhwgv" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.312544 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff-catalog-content\") pod \"certified-operators-nhwgv\" (UID: \"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff\") " pod="openshift-marketplace/certified-operators-nhwgv" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.312602 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:26 crc kubenswrapper[4802]: E1125 16:49:26.312973 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:26.812955944 +0000 UTC m=+149.957303130 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.413285 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.413523 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff-catalog-content\") pod \"certified-operators-nhwgv\" (UID: \"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff\") " pod="openshift-marketplace/certified-operators-nhwgv" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.413633 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff-utilities\") pod \"certified-operators-nhwgv\" (UID: \"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff\") " pod="openshift-marketplace/certified-operators-nhwgv" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.413685 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dvqh\" (UniqueName: \"kubernetes.io/projected/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff-kube-api-access-8dvqh\") pod \"certified-operators-nhwgv\" (UID: \"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff\") " pod="openshift-marketplace/certified-operators-nhwgv" Nov 25 16:49:26 crc kubenswrapper[4802]: E1125 16:49:26.414297 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:26.914266777 +0000 UTC m=+150.058613963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.414416 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff-catalog-content\") pod \"certified-operators-nhwgv\" (UID: \"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff\") " pod="openshift-marketplace/certified-operators-nhwgv" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.414703 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff-utilities\") pod \"certified-operators-nhwgv\" (UID: \"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff\") " pod="openshift-marketplace/certified-operators-nhwgv" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.437200 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jrq5g" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.437772 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rtgv8" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.449108 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dvqh\" (UniqueName: \"kubernetes.io/projected/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff-kube-api-access-8dvqh\") pod \"certified-operators-nhwgv\" (UID: \"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff\") " pod="openshift-marketplace/certified-operators-nhwgv" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.521925 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:26 crc kubenswrapper[4802]: E1125 16:49:26.522318 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:27.022300574 +0000 UTC m=+150.166647760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.523809 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nhwgv" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.574630 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r9nw2"] Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.622739 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:26 crc kubenswrapper[4802]: E1125 16:49:26.623156 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:27.123098374 +0000 UTC m=+150.267445560 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.724667 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:26 crc kubenswrapper[4802]: E1125 16:49:26.725068 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:27.225051635 +0000 UTC m=+150.369398821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.757529 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.822947 4802 patch_prober.go:28] interesting pod/router-default-5444994796-nr4zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 16:49:26 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Nov 25 16:49:26 crc kubenswrapper[4802]: [+]process-running ok Nov 25 16:49:26 crc kubenswrapper[4802]: healthz check failed Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.823245 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nr4zg" podUID="f7d5d65a-b6dd-45dd-9f35-5e02a915a64f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.825230 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:26 crc kubenswrapper[4802]: E1125 16:49:26.825654 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:27.32563384 +0000 UTC m=+150.469981026 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.846052 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"eb293a917e688e6f73a23449794f64069852fc7878a28af9f3ff61382100fbe6"} Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.855202 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9nw2" event={"ID":"9b19f412-72dd-4fe9-b0f1-3abc306653ae","Type":"ContainerStarted","Data":"f0614a9f2967c4648c9aa958d07ef7d86aef14fa6dbbee7ee8ddd5808040ec68"} Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.855242 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9nw2" event={"ID":"9b19f412-72dd-4fe9-b0f1-3abc306653ae","Type":"ContainerStarted","Data":"04f5263325584e02baed24b8a86b968ad3334c730b423247b61d707bffc44983"} Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.864946 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"74a06f10226c19a2c9e4a1a6707e8cf4e05a118c16404e0c3a97e2d94ed9b7c1"} Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.865813 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.872991 4802 generic.go:334] "Generic (PLEG): container finished" podID="17422b05-1782-421b-ab25-4ce61d267fc2" containerID="a686f2620d9bbe06f7f72afb13f5438c4d3ed7d2639d9a2f8dd8e80d01cb76a9" exitCode=0 Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.873075 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" event={"ID":"17422b05-1782-421b-ab25-4ce61d267fc2","Type":"ContainerDied","Data":"a686f2620d9bbe06f7f72afb13f5438c4d3ed7d2639d9a2f8dd8e80d01cb76a9"} Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.903089 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jrq5g"] Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.911953 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s684g" event={"ID":"d27bd203-adaf-4959-8907-416cd199df98","Type":"ContainerStarted","Data":"150e1098559399df5848fb4cb62e9ac3365a0a3d621205eba2816c8f0340b995"} Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.912010 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s684g" event={"ID":"d27bd203-adaf-4959-8907-416cd199df98","Type":"ContainerStarted","Data":"1e5a5bad4187de1bad77b46990712d4f301cacd7617f291f3f412e08b61ffe48"} Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.925646 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"5ff71ebcca5594fb2b935a59870fe9c6c2d841579ac8ad48a0565750baa35918"} Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.926755 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:26 crc kubenswrapper[4802]: I1125 16:49:26.929605 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rtgv8"] Nov 25 16:49:26 crc kubenswrapper[4802]: E1125 16:49:26.934207 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:27.434179219 +0000 UTC m=+150.578526405 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:26 crc kubenswrapper[4802]: W1125 16:49:26.969613 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod565aa173_3fac_4f4e_8fae_3fbe88184efb.slice/crio-404aeaf695d8748ce2de721578e021d59ff690576206fb5571ec41c5807bcbdb WatchSource:0}: Error finding container 404aeaf695d8748ce2de721578e021d59ff690576206fb5571ec41c5807bcbdb: Status 404 returned error can't find the container with id 404aeaf695d8748ce2de721578e021d59ff690576206fb5571ec41c5807bcbdb Nov 25 16:49:26 crc kubenswrapper[4802]: W1125 16:49:26.970353 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac5f001d_89fc_41e3_8a80_527e81eb55d0.slice/crio-d3addd95ffa846a7ed83c9b4e3f86402119f755d124ab1f869872b4272e9896e WatchSource:0}: Error finding container d3addd95ffa846a7ed83c9b4e3f86402119f755d124ab1f869872b4272e9896e: Status 404 returned error can't find the container with id d3addd95ffa846a7ed83c9b4e3f86402119f755d124ab1f869872b4272e9896e Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.028077 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:27 crc kubenswrapper[4802]: E1125 16:49:27.028227 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:27.528200917 +0000 UTC m=+150.672548103 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.029166 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:27 crc kubenswrapper[4802]: E1125 16:49:27.030935 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:27.53091819 +0000 UTC m=+150.675265376 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.110811 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 16:49:27 crc kubenswrapper[4802]: W1125 16:49:27.128068 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podcb067ba0_1c60_4b52_aa08_3675dd787fdf.slice/crio-0a94f7a480a4aaba35f7610a61efd2754df2c4dd668919b1fefec31da3416944 WatchSource:0}: Error finding container 0a94f7a480a4aaba35f7610a61efd2754df2c4dd668919b1fefec31da3416944: Status 404 returned error can't find the container with id 0a94f7a480a4aaba35f7610a61efd2754df2c4dd668919b1fefec31da3416944 Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.135008 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:27 crc kubenswrapper[4802]: E1125 16:49:27.135421 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:27.635361839 +0000 UTC m=+150.779709025 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.144734 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nhwgv"] Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.175929 4802 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.236741 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:27 crc kubenswrapper[4802]: E1125 16:49:27.237215 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:27.737193937 +0000 UTC m=+150.881541193 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.338222 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:27 crc kubenswrapper[4802]: E1125 16:49:27.338716 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:27.838697196 +0000 UTC m=+150.983044392 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.440149 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:27 crc kubenswrapper[4802]: E1125 16:49:27.440563 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:27.940543905 +0000 UTC m=+151.084891091 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.541968 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:27 crc kubenswrapper[4802]: E1125 16:49:27.542108 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:28.042085696 +0000 UTC m=+151.186432882 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.542239 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:27 crc kubenswrapper[4802]: E1125 16:49:27.542578 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:28.042565818 +0000 UTC m=+151.186913084 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.564360 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.564447 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.569755 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-469c8"] Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.572371 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-469c8" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.573466 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.575247 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.588332 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-469c8"] Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.591972 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.592020 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.606591 4802 patch_prober.go:28] interesting pod/apiserver-76f77b778f-hzx2q container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 25 16:49:27 crc kubenswrapper[4802]: [+]log ok Nov 25 16:49:27 crc kubenswrapper[4802]: [+]etcd ok Nov 25 16:49:27 crc kubenswrapper[4802]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 25 16:49:27 crc kubenswrapper[4802]: [+]poststarthook/generic-apiserver-start-informers ok Nov 25 16:49:27 crc kubenswrapper[4802]: [+]poststarthook/max-in-flight-filter ok Nov 25 16:49:27 crc kubenswrapper[4802]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 25 16:49:27 crc kubenswrapper[4802]: [+]poststarthook/image.openshift.io-apiserver-caches ok Nov 25 16:49:27 crc kubenswrapper[4802]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Nov 25 16:49:27 crc kubenswrapper[4802]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Nov 25 16:49:27 crc kubenswrapper[4802]: [+]poststarthook/project.openshift.io-projectcache ok Nov 25 16:49:27 crc kubenswrapper[4802]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Nov 25 16:49:27 crc kubenswrapper[4802]: [+]poststarthook/openshift.io-startinformers ok Nov 25 16:49:27 crc kubenswrapper[4802]: [+]poststarthook/openshift.io-restmapperupdater ok Nov 25 16:49:27 crc kubenswrapper[4802]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 25 16:49:27 crc kubenswrapper[4802]: livez check failed Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.606674 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" podUID="a0a85d28-62a0-4ddc-ab49-c73b6db760a0" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.643246 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:27 crc kubenswrapper[4802]: E1125 16:49:27.643541 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 16:49:28.143519873 +0000 UTC m=+151.287867059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.643636 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:27 crc kubenswrapper[4802]: E1125 16:49:27.643980 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 16:49:28.143966875 +0000 UTC m=+151.288314061 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kst7h" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.655266 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.683285 4802 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-25T16:49:27.175948798Z","Handler":null,"Name":""} Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.696212 4802 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.696252 4802 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.746413 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.746965 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr8ld\" (UniqueName: \"kubernetes.io/projected/8790f862-7ccc-4b14-8c99-1f64349079e0-kube-api-access-sr8ld\") pod \"redhat-marketplace-469c8\" (UID: \"8790f862-7ccc-4b14-8c99-1f64349079e0\") " pod="openshift-marketplace/redhat-marketplace-469c8" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.747103 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8790f862-7ccc-4b14-8c99-1f64349079e0-utilities\") pod \"redhat-marketplace-469c8\" (UID: \"8790f862-7ccc-4b14-8c99-1f64349079e0\") " pod="openshift-marketplace/redhat-marketplace-469c8" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.747407 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8790f862-7ccc-4b14-8c99-1f64349079e0-catalog-content\") pod \"redhat-marketplace-469c8\" (UID: \"8790f862-7ccc-4b14-8c99-1f64349079e0\") " pod="openshift-marketplace/redhat-marketplace-469c8" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.751358 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.813329 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.816767 4802 patch_prober.go:28] interesting pod/router-default-5444994796-nr4zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 16:49:27 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Nov 25 16:49:27 crc kubenswrapper[4802]: [+]process-running ok Nov 25 16:49:27 crc kubenswrapper[4802]: healthz check failed Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.816873 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nr4zg" podUID="f7d5d65a-b6dd-45dd-9f35-5e02a915a64f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.848590 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8790f862-7ccc-4b14-8c99-1f64349079e0-catalog-content\") pod \"redhat-marketplace-469c8\" (UID: \"8790f862-7ccc-4b14-8c99-1f64349079e0\") " pod="openshift-marketplace/redhat-marketplace-469c8" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.848700 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr8ld\" (UniqueName: \"kubernetes.io/projected/8790f862-7ccc-4b14-8c99-1f64349079e0-kube-api-access-sr8ld\") pod \"redhat-marketplace-469c8\" (UID: \"8790f862-7ccc-4b14-8c99-1f64349079e0\") " pod="openshift-marketplace/redhat-marketplace-469c8" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.848747 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8790f862-7ccc-4b14-8c99-1f64349079e0-utilities\") pod \"redhat-marketplace-469c8\" (UID: \"8790f862-7ccc-4b14-8c99-1f64349079e0\") " pod="openshift-marketplace/redhat-marketplace-469c8" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.848781 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.849397 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8790f862-7ccc-4b14-8c99-1f64349079e0-catalog-content\") pod \"redhat-marketplace-469c8\" (UID: \"8790f862-7ccc-4b14-8c99-1f64349079e0\") " pod="openshift-marketplace/redhat-marketplace-469c8" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.849775 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8790f862-7ccc-4b14-8c99-1f64349079e0-utilities\") pod \"redhat-marketplace-469c8\" (UID: \"8790f862-7ccc-4b14-8c99-1f64349079e0\") " pod="openshift-marketplace/redhat-marketplace-469c8" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.873788 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr8ld\" (UniqueName: \"kubernetes.io/projected/8790f862-7ccc-4b14-8c99-1f64349079e0-kube-api-access-sr8ld\") pod \"redhat-marketplace-469c8\" (UID: \"8790f862-7ccc-4b14-8c99-1f64349079e0\") " pod="openshift-marketplace/redhat-marketplace-469c8" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.882868 4802 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.882958 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.893073 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-469c8" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.913669 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kst7h\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.947597 4802 patch_prober.go:28] interesting pod/downloads-7954f5f757-z8dgw container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.947648 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-z8dgw" podUID="8a3ebc4f-0491-4dc1-8c5d-adb851b6ef87" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.947886 4802 patch_prober.go:28] interesting pod/downloads-7954f5f757-z8dgw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.947902 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-z8dgw" podUID="8a3ebc4f-0491-4dc1-8c5d-adb851b6ef87" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.947998 4802 generic.go:334] "Generic (PLEG): container finished" podID="ac5f001d-89fc-41e3-8a80-527e81eb55d0" containerID="14ea03ef1504d38cc6ea45818f23bd0acf170243150a6f8f5c15f6b3bbe81f55" exitCode=0 Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.948082 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jrq5g" event={"ID":"ac5f001d-89fc-41e3-8a80-527e81eb55d0","Type":"ContainerDied","Data":"14ea03ef1504d38cc6ea45818f23bd0acf170243150a6f8f5c15f6b3bbe81f55"} Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.948147 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jrq5g" event={"ID":"ac5f001d-89fc-41e3-8a80-527e81eb55d0","Type":"ContainerStarted","Data":"d3addd95ffa846a7ed83c9b4e3f86402119f755d124ab1f869872b4272e9896e"} Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.953427 4802 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.957297 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rtgv8" event={"ID":"565aa173-3fac-4f4e-8fae-3fbe88184efb","Type":"ContainerDied","Data":"7c845f5220c08bf1df0f8ed5cec4f02229391b92ad53c79d8566ce5b1b3c262e"} Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.957189 4802 generic.go:334] "Generic (PLEG): container finished" podID="565aa173-3fac-4f4e-8fae-3fbe88184efb" containerID="7c845f5220c08bf1df0f8ed5cec4f02229391b92ad53c79d8566ce5b1b3c262e" exitCode=0 Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.959805 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rtgv8" event={"ID":"565aa173-3fac-4f4e-8fae-3fbe88184efb","Type":"ContainerStarted","Data":"404aeaf695d8748ce2de721578e021d59ff690576206fb5571ec41c5807bcbdb"} Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.970459 4802 generic.go:334] "Generic (PLEG): container finished" podID="9b19f412-72dd-4fe9-b0f1-3abc306653ae" containerID="f0614a9f2967c4648c9aa958d07ef7d86aef14fa6dbbee7ee8ddd5808040ec68" exitCode=0 Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.970637 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9nw2" event={"ID":"9b19f412-72dd-4fe9-b0f1-3abc306653ae","Type":"ContainerDied","Data":"f0614a9f2967c4648c9aa958d07ef7d86aef14fa6dbbee7ee8ddd5808040ec68"} Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.983338 4802 generic.go:334] "Generic (PLEG): container finished" podID="078b7cce-5ee4-4c6e-9d9b-5332a6f070ff" containerID="750e351dc53ca74511b6d49c9abfc2fbfb9853bb6975c10e76703d7a1fc2d106" exitCode=0 Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.983468 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhwgv" event={"ID":"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff","Type":"ContainerDied","Data":"750e351dc53ca74511b6d49c9abfc2fbfb9853bb6975c10e76703d7a1fc2d106"} Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.983494 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhwgv" event={"ID":"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff","Type":"ContainerStarted","Data":"36dfce8c051a60cff6c7ef28623eff7b45d14383b43e4820f4ae1adec82a8227"} Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.989080 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gvlq8"] Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.990395 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gvlq8" Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.998478 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s684g" event={"ID":"d27bd203-adaf-4959-8907-416cd199df98","Type":"ContainerStarted","Data":"f848e58031417b9dc7737b313cdf95b1162994a0eec7e25371def2e1bde14a7c"} Nov 25 16:49:27 crc kubenswrapper[4802]: I1125 16:49:27.998952 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gvlq8"] Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.003722 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.004969 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.010749 4802 patch_prober.go:28] interesting pod/console-f9d7485db-p7zvz container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.010810 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-p7zvz" podUID="df33868c-7e7c-4d67-b478-40ee1d7ef69d" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.024365 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"cb067ba0-1c60-4b52-aa08-3675dd787fdf","Type":"ContainerStarted","Data":"6f93e0083406870f1449cd132eb198346144ad24810c1b52b838359026b1a3cf"} Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.024463 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"cb067ba0-1c60-4b52-aa08-3675dd787fdf","Type":"ContainerStarted","Data":"0a94f7a480a4aaba35f7610a61efd2754df2c4dd668919b1fefec31da3416944"} Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.024915 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.032350 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hljn2" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.114049 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-s684g" podStartSLOduration=13.114011486 podStartE2EDuration="13.114011486s" podCreationTimestamp="2025-11-25 16:49:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:28.083352466 +0000 UTC m=+151.227699662" watchObservedRunningTime="2025-11-25 16:49:28.114011486 +0000 UTC m=+151.258358672" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.136437 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.136262069 podStartE2EDuration="3.136262069s" podCreationTimestamp="2025-11-25 16:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:28.13443467 +0000 UTC m=+151.278781856" watchObservedRunningTime="2025-11-25 16:49:28.136262069 +0000 UTC m=+151.280609255" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.179694 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbnvk\" (UniqueName: \"kubernetes.io/projected/ce7b5945-011b-46bb-9af5-96fe9d8e0f89-kube-api-access-kbnvk\") pod \"redhat-marketplace-gvlq8\" (UID: \"ce7b5945-011b-46bb-9af5-96fe9d8e0f89\") " pod="openshift-marketplace/redhat-marketplace-gvlq8" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.179811 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce7b5945-011b-46bb-9af5-96fe9d8e0f89-catalog-content\") pod \"redhat-marketplace-gvlq8\" (UID: \"ce7b5945-011b-46bb-9af5-96fe9d8e0f89\") " pod="openshift-marketplace/redhat-marketplace-gvlq8" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.180525 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce7b5945-011b-46bb-9af5-96fe9d8e0f89-utilities\") pod \"redhat-marketplace-gvlq8\" (UID: \"ce7b5945-011b-46bb-9af5-96fe9d8e0f89\") " pod="openshift-marketplace/redhat-marketplace-gvlq8" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.281595 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce7b5945-011b-46bb-9af5-96fe9d8e0f89-utilities\") pod \"redhat-marketplace-gvlq8\" (UID: \"ce7b5945-011b-46bb-9af5-96fe9d8e0f89\") " pod="openshift-marketplace/redhat-marketplace-gvlq8" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.281669 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbnvk\" (UniqueName: \"kubernetes.io/projected/ce7b5945-011b-46bb-9af5-96fe9d8e0f89-kube-api-access-kbnvk\") pod \"redhat-marketplace-gvlq8\" (UID: \"ce7b5945-011b-46bb-9af5-96fe9d8e0f89\") " pod="openshift-marketplace/redhat-marketplace-gvlq8" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.281698 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce7b5945-011b-46bb-9af5-96fe9d8e0f89-catalog-content\") pod \"redhat-marketplace-gvlq8\" (UID: \"ce7b5945-011b-46bb-9af5-96fe9d8e0f89\") " pod="openshift-marketplace/redhat-marketplace-gvlq8" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.282149 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce7b5945-011b-46bb-9af5-96fe9d8e0f89-catalog-content\") pod \"redhat-marketplace-gvlq8\" (UID: \"ce7b5945-011b-46bb-9af5-96fe9d8e0f89\") " pod="openshift-marketplace/redhat-marketplace-gvlq8" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.282356 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce7b5945-011b-46bb-9af5-96fe9d8e0f89-utilities\") pod \"redhat-marketplace-gvlq8\" (UID: \"ce7b5945-011b-46bb-9af5-96fe9d8e0f89\") " pod="openshift-marketplace/redhat-marketplace-gvlq8" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.309692 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.311521 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hb9lv" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.328277 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbnvk\" (UniqueName: \"kubernetes.io/projected/ce7b5945-011b-46bb-9af5-96fe9d8e0f89-kube-api-access-kbnvk\") pod \"redhat-marketplace-gvlq8\" (UID: \"ce7b5945-011b-46bb-9af5-96fe9d8e0f89\") " pod="openshift-marketplace/redhat-marketplace-gvlq8" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.401264 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-469c8"] Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.432636 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.457159 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kst7h"] Nov 25 16:49:28 crc kubenswrapper[4802]: W1125 16:49:28.481525 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c5d984b_9b31_4997_904a_f27989f78a8c.slice/crio-0d6e24976433a7ec3a792489e3bf828d6565a5e2d98a176a1cab843c15c9c970 WatchSource:0}: Error finding container 0d6e24976433a7ec3a792489e3bf828d6565a5e2d98a176a1cab843c15c9c970: Status 404 returned error can't find the container with id 0d6e24976433a7ec3a792489e3bf828d6565a5e2d98a176a1cab843c15c9c970 Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.496375 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/17422b05-1782-421b-ab25-4ce61d267fc2-config-volume\") pod \"17422b05-1782-421b-ab25-4ce61d267fc2\" (UID: \"17422b05-1782-421b-ab25-4ce61d267fc2\") " Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.496511 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/17422b05-1782-421b-ab25-4ce61d267fc2-secret-volume\") pod \"17422b05-1782-421b-ab25-4ce61d267fc2\" (UID: \"17422b05-1782-421b-ab25-4ce61d267fc2\") " Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.496565 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t96bv\" (UniqueName: \"kubernetes.io/projected/17422b05-1782-421b-ab25-4ce61d267fc2-kube-api-access-t96bv\") pod \"17422b05-1782-421b-ab25-4ce61d267fc2\" (UID: \"17422b05-1782-421b-ab25-4ce61d267fc2\") " Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.497190 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17422b05-1782-421b-ab25-4ce61d267fc2-config-volume" (OuterVolumeSpecName: "config-volume") pod "17422b05-1782-421b-ab25-4ce61d267fc2" (UID: "17422b05-1782-421b-ab25-4ce61d267fc2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.502254 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17422b05-1782-421b-ab25-4ce61d267fc2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "17422b05-1782-421b-ab25-4ce61d267fc2" (UID: "17422b05-1782-421b-ab25-4ce61d267fc2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.503490 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17422b05-1782-421b-ab25-4ce61d267fc2-kube-api-access-t96bv" (OuterVolumeSpecName: "kube-api-access-t96bv") pod "17422b05-1782-421b-ab25-4ce61d267fc2" (UID: "17422b05-1782-421b-ab25-4ce61d267fc2"). InnerVolumeSpecName "kube-api-access-t96bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.518189 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-nmzxv" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.593656 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29vlk" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.598820 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t96bv\" (UniqueName: \"kubernetes.io/projected/17422b05-1782-421b-ab25-4ce61d267fc2-kube-api-access-t96bv\") on node \"crc\" DevicePath \"\"" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.598970 4802 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/17422b05-1782-421b-ab25-4ce61d267fc2-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.599397 4802 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/17422b05-1782-421b-ab25-4ce61d267fc2-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.622525 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gvlq8" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.818649 4802 patch_prober.go:28] interesting pod/router-default-5444994796-nr4zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 16:49:28 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Nov 25 16:49:28 crc kubenswrapper[4802]: [+]process-running ok Nov 25 16:49:28 crc kubenswrapper[4802]: healthz check failed Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.819283 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nr4zg" podUID="f7d5d65a-b6dd-45dd-9f35-5e02a915a64f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.837824 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gvlq8"] Nov 25 16:49:28 crc kubenswrapper[4802]: W1125 16:49:28.847678 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce7b5945_011b_46bb_9af5_96fe9d8e0f89.slice/crio-b50e23a0369c8f32298eb164697f2220bc192cdb1613afe62311e1f17197165d WatchSource:0}: Error finding container b50e23a0369c8f32298eb164697f2220bc192cdb1613afe62311e1f17197165d: Status 404 returned error can't find the container with id b50e23a0369c8f32298eb164697f2220bc192cdb1613afe62311e1f17197165d Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.980059 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wsq47"] Nov 25 16:49:28 crc kubenswrapper[4802]: E1125 16:49:28.980429 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17422b05-1782-421b-ab25-4ce61d267fc2" containerName="collect-profiles" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.980445 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="17422b05-1782-421b-ab25-4ce61d267fc2" containerName="collect-profiles" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.980574 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="17422b05-1782-421b-ab25-4ce61d267fc2" containerName="collect-profiles" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.981498 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wsq47" Nov 25 16:49:28 crc kubenswrapper[4802]: I1125 16:49:28.987092 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.000383 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wsq47"] Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.004367 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21-catalog-content\") pod \"redhat-operators-wsq47\" (UID: \"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21\") " pod="openshift-marketplace/redhat-operators-wsq47" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.004455 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21-utilities\") pod \"redhat-operators-wsq47\" (UID: \"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21\") " pod="openshift-marketplace/redhat-operators-wsq47" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.004479 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ckk8\" (UniqueName: \"kubernetes.io/projected/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21-kube-api-access-4ckk8\") pod \"redhat-operators-wsq47\" (UID: \"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21\") " pod="openshift-marketplace/redhat-operators-wsq47" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.043288 4802 generic.go:334] "Generic (PLEG): container finished" podID="8790f862-7ccc-4b14-8c99-1f64349079e0" containerID="d413871f5ffe933b3d095fd476deac62abe1f70847ee024955eae178d759f358" exitCode=0 Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.043382 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-469c8" event={"ID":"8790f862-7ccc-4b14-8c99-1f64349079e0","Type":"ContainerDied","Data":"d413871f5ffe933b3d095fd476deac62abe1f70847ee024955eae178d759f358"} Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.043423 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-469c8" event={"ID":"8790f862-7ccc-4b14-8c99-1f64349079e0","Type":"ContainerStarted","Data":"3102be8c615409385fbccc78cdee471dd76eb9e6fbfa1210f19d86d6e87feaef"} Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.048098 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" event={"ID":"17422b05-1782-421b-ab25-4ce61d267fc2","Type":"ContainerDied","Data":"43da95e700a9b1716cfbf18b8b617d01f07c43141ecb1458fe75d8650f2dd033"} Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.048163 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43da95e700a9b1716cfbf18b8b617d01f07c43141ecb1458fe75d8650f2dd033" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.048236 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401485-28zdh" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.067231 4802 generic.go:334] "Generic (PLEG): container finished" podID="cb067ba0-1c60-4b52-aa08-3675dd787fdf" containerID="6f93e0083406870f1449cd132eb198346144ad24810c1b52b838359026b1a3cf" exitCode=0 Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.067376 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"cb067ba0-1c60-4b52-aa08-3675dd787fdf","Type":"ContainerDied","Data":"6f93e0083406870f1449cd132eb198346144ad24810c1b52b838359026b1a3cf"} Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.078440 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gvlq8" event={"ID":"ce7b5945-011b-46bb-9af5-96fe9d8e0f89","Type":"ContainerStarted","Data":"b50e23a0369c8f32298eb164697f2220bc192cdb1613afe62311e1f17197165d"} Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.088645 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" event={"ID":"8c5d984b-9b31-4997-904a-f27989f78a8c","Type":"ContainerStarted","Data":"247f6204a5cf8a2e11a741690fae81646177dc55c4081145cf72376b1e6e25c3"} Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.088752 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.088775 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" event={"ID":"8c5d984b-9b31-4997-904a-f27989f78a8c","Type":"ContainerStarted","Data":"0d6e24976433a7ec3a792489e3bf828d6565a5e2d98a176a1cab843c15c9c970"} Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.114186 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21-catalog-content\") pod \"redhat-operators-wsq47\" (UID: \"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21\") " pod="openshift-marketplace/redhat-operators-wsq47" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.114354 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21-utilities\") pod \"redhat-operators-wsq47\" (UID: \"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21\") " pod="openshift-marketplace/redhat-operators-wsq47" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.114376 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ckk8\" (UniqueName: \"kubernetes.io/projected/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21-kube-api-access-4ckk8\") pod \"redhat-operators-wsq47\" (UID: \"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21\") " pod="openshift-marketplace/redhat-operators-wsq47" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.116544 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21-catalog-content\") pod \"redhat-operators-wsq47\" (UID: \"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21\") " pod="openshift-marketplace/redhat-operators-wsq47" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.117411 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21-utilities\") pod \"redhat-operators-wsq47\" (UID: \"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21\") " pod="openshift-marketplace/redhat-operators-wsq47" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.142826 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ckk8\" (UniqueName: \"kubernetes.io/projected/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21-kube-api-access-4ckk8\") pod \"redhat-operators-wsq47\" (UID: \"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21\") " pod="openshift-marketplace/redhat-operators-wsq47" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.153555 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" podStartSLOduration=131.153539382 podStartE2EDuration="2m11.153539382s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:29.151546958 +0000 UTC m=+152.295894134" watchObservedRunningTime="2025-11-25 16:49:29.153539382 +0000 UTC m=+152.297886568" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.303379 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wsq47" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.386551 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t56jf"] Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.387965 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t56jf" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.394139 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t56jf"] Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.418540 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31e04778-8768-4164-a76d-8afc694db820-utilities\") pod \"redhat-operators-t56jf\" (UID: \"31e04778-8768-4164-a76d-8afc694db820\") " pod="openshift-marketplace/redhat-operators-t56jf" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.418625 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31e04778-8768-4164-a76d-8afc694db820-catalog-content\") pod \"redhat-operators-t56jf\" (UID: \"31e04778-8768-4164-a76d-8afc694db820\") " pod="openshift-marketplace/redhat-operators-t56jf" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.418742 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq2wd\" (UniqueName: \"kubernetes.io/projected/31e04778-8768-4164-a76d-8afc694db820-kube-api-access-cq2wd\") pod \"redhat-operators-t56jf\" (UID: \"31e04778-8768-4164-a76d-8afc694db820\") " pod="openshift-marketplace/redhat-operators-t56jf" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.514262 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.519871 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq2wd\" (UniqueName: \"kubernetes.io/projected/31e04778-8768-4164-a76d-8afc694db820-kube-api-access-cq2wd\") pod \"redhat-operators-t56jf\" (UID: \"31e04778-8768-4164-a76d-8afc694db820\") " pod="openshift-marketplace/redhat-operators-t56jf" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.519967 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31e04778-8768-4164-a76d-8afc694db820-utilities\") pod \"redhat-operators-t56jf\" (UID: \"31e04778-8768-4164-a76d-8afc694db820\") " pod="openshift-marketplace/redhat-operators-t56jf" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.520019 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31e04778-8768-4164-a76d-8afc694db820-catalog-content\") pod \"redhat-operators-t56jf\" (UID: \"31e04778-8768-4164-a76d-8afc694db820\") " pod="openshift-marketplace/redhat-operators-t56jf" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.520846 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31e04778-8768-4164-a76d-8afc694db820-catalog-content\") pod \"redhat-operators-t56jf\" (UID: \"31e04778-8768-4164-a76d-8afc694db820\") " pod="openshift-marketplace/redhat-operators-t56jf" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.521089 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31e04778-8768-4164-a76d-8afc694db820-utilities\") pod \"redhat-operators-t56jf\" (UID: \"31e04778-8768-4164-a76d-8afc694db820\") " pod="openshift-marketplace/redhat-operators-t56jf" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.561200 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq2wd\" (UniqueName: \"kubernetes.io/projected/31e04778-8768-4164-a76d-8afc694db820-kube-api-access-cq2wd\") pod \"redhat-operators-t56jf\" (UID: \"31e04778-8768-4164-a76d-8afc694db820\") " pod="openshift-marketplace/redhat-operators-t56jf" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.724752 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t56jf" Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.774280 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wsq47"] Nov 25 16:49:29 crc kubenswrapper[4802]: W1125 16:49:29.790233 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55ae2bee_8b1d_46a6_a6c2_76c97f8bdf21.slice/crio-cd0e0ac1919ff451dfe62d09dd9b8169eecda46a1312f0a13824aa8ca96a8aba WatchSource:0}: Error finding container cd0e0ac1919ff451dfe62d09dd9b8169eecda46a1312f0a13824aa8ca96a8aba: Status 404 returned error can't find the container with id cd0e0ac1919ff451dfe62d09dd9b8169eecda46a1312f0a13824aa8ca96a8aba Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.818178 4802 patch_prober.go:28] interesting pod/router-default-5444994796-nr4zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 16:49:29 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Nov 25 16:49:29 crc kubenswrapper[4802]: [+]process-running ok Nov 25 16:49:29 crc kubenswrapper[4802]: healthz check failed Nov 25 16:49:29 crc kubenswrapper[4802]: I1125 16:49:29.818266 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nr4zg" podUID="f7d5d65a-b6dd-45dd-9f35-5e02a915a64f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.118452 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsq47" event={"ID":"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21","Type":"ContainerStarted","Data":"cd0e0ac1919ff451dfe62d09dd9b8169eecda46a1312f0a13824aa8ca96a8aba"} Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.122147 4802 generic.go:334] "Generic (PLEG): container finished" podID="ce7b5945-011b-46bb-9af5-96fe9d8e0f89" containerID="c018dfd930fd159185fa7f25c5a21a7b4ab38ee055480cfb4428ae5efc61bd97" exitCode=0 Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.126796 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gvlq8" event={"ID":"ce7b5945-011b-46bb-9af5-96fe9d8e0f89","Type":"ContainerDied","Data":"c018dfd930fd159185fa7f25c5a21a7b4ab38ee055480cfb4428ae5efc61bd97"} Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.278286 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t56jf"] Nov 25 16:49:30 crc kubenswrapper[4802]: W1125 16:49:30.328545 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31e04778_8768_4164_a76d_8afc694db820.slice/crio-25a696ad4e97c044c3dd6ee921273d09badc0af503405e5a6f92b49217e0e5d3 WatchSource:0}: Error finding container 25a696ad4e97c044c3dd6ee921273d09badc0af503405e5a6f92b49217e0e5d3: Status 404 returned error can't find the container with id 25a696ad4e97c044c3dd6ee921273d09badc0af503405e5a6f92b49217e0e5d3 Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.562630 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.691052 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 16:49:30 crc kubenswrapper[4802]: E1125 16:49:30.691351 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb067ba0-1c60-4b52-aa08-3675dd787fdf" containerName="pruner" Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.696321 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb067ba0-1c60-4b52-aa08-3675dd787fdf" containerName="pruner" Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.696658 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb067ba0-1c60-4b52-aa08-3675dd787fdf" containerName="pruner" Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.697197 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.697302 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.700517 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.700677 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.742750 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cb067ba0-1c60-4b52-aa08-3675dd787fdf-kubelet-dir\") pod \"cb067ba0-1c60-4b52-aa08-3675dd787fdf\" (UID: \"cb067ba0-1c60-4b52-aa08-3675dd787fdf\") " Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.742845 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb067ba0-1c60-4b52-aa08-3675dd787fdf-kube-api-access\") pod \"cb067ba0-1c60-4b52-aa08-3675dd787fdf\" (UID: \"cb067ba0-1c60-4b52-aa08-3675dd787fdf\") " Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.743315 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb067ba0-1c60-4b52-aa08-3675dd787fdf-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "cb067ba0-1c60-4b52-aa08-3675dd787fdf" (UID: "cb067ba0-1c60-4b52-aa08-3675dd787fdf"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.755551 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb067ba0-1c60-4b52-aa08-3675dd787fdf-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "cb067ba0-1c60-4b52-aa08-3675dd787fdf" (UID: "cb067ba0-1c60-4b52-aa08-3675dd787fdf"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.817248 4802 patch_prober.go:28] interesting pod/router-default-5444994796-nr4zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 16:49:30 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Nov 25 16:49:30 crc kubenswrapper[4802]: [+]process-running ok Nov 25 16:49:30 crc kubenswrapper[4802]: healthz check failed Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.817320 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nr4zg" podUID="f7d5d65a-b6dd-45dd-9f35-5e02a915a64f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.844259 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/73a060c4-b1a7-4d9b-875c-e0c426e2fa4f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"73a060c4-b1a7-4d9b-875c-e0c426e2fa4f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.845175 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/73a060c4-b1a7-4d9b-875c-e0c426e2fa4f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"73a060c4-b1a7-4d9b-875c-e0c426e2fa4f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.845258 4802 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cb067ba0-1c60-4b52-aa08-3675dd787fdf-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.845271 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb067ba0-1c60-4b52-aa08-3675dd787fdf-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.946606 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/73a060c4-b1a7-4d9b-875c-e0c426e2fa4f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"73a060c4-b1a7-4d9b-875c-e0c426e2fa4f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.946703 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/73a060c4-b1a7-4d9b-875c-e0c426e2fa4f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"73a060c4-b1a7-4d9b-875c-e0c426e2fa4f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.946820 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/73a060c4-b1a7-4d9b-875c-e0c426e2fa4f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"73a060c4-b1a7-4d9b-875c-e0c426e2fa4f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 16:49:30 crc kubenswrapper[4802]: I1125 16:49:30.963373 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/73a060c4-b1a7-4d9b-875c-e0c426e2fa4f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"73a060c4-b1a7-4d9b-875c-e0c426e2fa4f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 16:49:31 crc kubenswrapper[4802]: I1125 16:49:31.070910 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 16:49:31 crc kubenswrapper[4802]: I1125 16:49:31.138729 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 16:49:31 crc kubenswrapper[4802]: I1125 16:49:31.138742 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"cb067ba0-1c60-4b52-aa08-3675dd787fdf","Type":"ContainerDied","Data":"0a94f7a480a4aaba35f7610a61efd2754df2c4dd668919b1fefec31da3416944"} Nov 25 16:49:31 crc kubenswrapper[4802]: I1125 16:49:31.138777 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a94f7a480a4aaba35f7610a61efd2754df2c4dd668919b1fefec31da3416944" Nov 25 16:49:31 crc kubenswrapper[4802]: I1125 16:49:31.143944 4802 generic.go:334] "Generic (PLEG): container finished" podID="55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21" containerID="ed1e7d350328f21c0c42f7a18af9596b5e7fcc77c2e114918e9152d87ff88c3a" exitCode=0 Nov 25 16:49:31 crc kubenswrapper[4802]: I1125 16:49:31.144048 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsq47" event={"ID":"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21","Type":"ContainerDied","Data":"ed1e7d350328f21c0c42f7a18af9596b5e7fcc77c2e114918e9152d87ff88c3a"} Nov 25 16:49:31 crc kubenswrapper[4802]: I1125 16:49:31.149825 4802 generic.go:334] "Generic (PLEG): container finished" podID="31e04778-8768-4164-a76d-8afc694db820" containerID="bfa5d983d60beb0de2d7523d8ac39782d9237fdee78e30471514de377d6e5505" exitCode=0 Nov 25 16:49:31 crc kubenswrapper[4802]: I1125 16:49:31.150236 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t56jf" event={"ID":"31e04778-8768-4164-a76d-8afc694db820","Type":"ContainerDied","Data":"bfa5d983d60beb0de2d7523d8ac39782d9237fdee78e30471514de377d6e5505"} Nov 25 16:49:31 crc kubenswrapper[4802]: I1125 16:49:31.150409 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t56jf" event={"ID":"31e04778-8768-4164-a76d-8afc694db820","Type":"ContainerStarted","Data":"25a696ad4e97c044c3dd6ee921273d09badc0af503405e5a6f92b49217e0e5d3"} Nov 25 16:49:31 crc kubenswrapper[4802]: I1125 16:49:31.493076 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 16:49:31 crc kubenswrapper[4802]: W1125 16:49:31.502066 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod73a060c4_b1a7_4d9b_875c_e0c426e2fa4f.slice/crio-960e8ac4d7ab14413d4cef40cc5a19da7db93d71741190f17d9a114a5788dfdb WatchSource:0}: Error finding container 960e8ac4d7ab14413d4cef40cc5a19da7db93d71741190f17d9a114a5788dfdb: Status 404 returned error can't find the container with id 960e8ac4d7ab14413d4cef40cc5a19da7db93d71741190f17d9a114a5788dfdb Nov 25 16:49:31 crc kubenswrapper[4802]: I1125 16:49:31.816490 4802 patch_prober.go:28] interesting pod/router-default-5444994796-nr4zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 16:49:31 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Nov 25 16:49:31 crc kubenswrapper[4802]: [+]process-running ok Nov 25 16:49:31 crc kubenswrapper[4802]: healthz check failed Nov 25 16:49:31 crc kubenswrapper[4802]: I1125 16:49:31.817051 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nr4zg" podUID="f7d5d65a-b6dd-45dd-9f35-5e02a915a64f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 16:49:32 crc kubenswrapper[4802]: I1125 16:49:32.164606 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"73a060c4-b1a7-4d9b-875c-e0c426e2fa4f","Type":"ContainerStarted","Data":"960e8ac4d7ab14413d4cef40cc5a19da7db93d71741190f17d9a114a5788dfdb"} Nov 25 16:49:32 crc kubenswrapper[4802]: I1125 16:49:32.599702 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:32 crc kubenswrapper[4802]: I1125 16:49:32.605695 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-hzx2q" Nov 25 16:49:32 crc kubenswrapper[4802]: I1125 16:49:32.817096 4802 patch_prober.go:28] interesting pod/router-default-5444994796-nr4zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 16:49:32 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Nov 25 16:49:32 crc kubenswrapper[4802]: [+]process-running ok Nov 25 16:49:32 crc kubenswrapper[4802]: healthz check failed Nov 25 16:49:32 crc kubenswrapper[4802]: I1125 16:49:32.817175 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nr4zg" podUID="f7d5d65a-b6dd-45dd-9f35-5e02a915a64f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 16:49:33 crc kubenswrapper[4802]: I1125 16:49:33.182621 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"73a060c4-b1a7-4d9b-875c-e0c426e2fa4f","Type":"ContainerStarted","Data":"4d2584388d7587772417143b3849be541b6353df01a5adbb8e0a54b7cad8989d"} Nov 25 16:49:33 crc kubenswrapper[4802]: I1125 16:49:33.211164 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.211139803 podStartE2EDuration="3.211139803s" podCreationTimestamp="2025-11-25 16:49:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:49:33.205773009 +0000 UTC m=+156.350120195" watchObservedRunningTime="2025-11-25 16:49:33.211139803 +0000 UTC m=+156.355486989" Nov 25 16:49:33 crc kubenswrapper[4802]: I1125 16:49:33.616312 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-v6gj4" Nov 25 16:49:33 crc kubenswrapper[4802]: I1125 16:49:33.816281 4802 patch_prober.go:28] interesting pod/router-default-5444994796-nr4zg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 16:49:33 crc kubenswrapper[4802]: [-]has-synced failed: reason withheld Nov 25 16:49:33 crc kubenswrapper[4802]: [+]process-running ok Nov 25 16:49:33 crc kubenswrapper[4802]: healthz check failed Nov 25 16:49:33 crc kubenswrapper[4802]: I1125 16:49:33.816378 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-nr4zg" podUID="f7d5d65a-b6dd-45dd-9f35-5e02a915a64f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 16:49:34 crc kubenswrapper[4802]: I1125 16:49:34.818604 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:34 crc kubenswrapper[4802]: I1125 16:49:34.823523 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-nr4zg" Nov 25 16:49:35 crc kubenswrapper[4802]: I1125 16:49:35.207749 4802 generic.go:334] "Generic (PLEG): container finished" podID="73a060c4-b1a7-4d9b-875c-e0c426e2fa4f" containerID="4d2584388d7587772417143b3849be541b6353df01a5adbb8e0a54b7cad8989d" exitCode=0 Nov 25 16:49:35 crc kubenswrapper[4802]: I1125 16:49:35.207784 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"73a060c4-b1a7-4d9b-875c-e0c426e2fa4f","Type":"ContainerDied","Data":"4d2584388d7587772417143b3849be541b6353df01a5adbb8e0a54b7cad8989d"} Nov 25 16:49:37 crc kubenswrapper[4802]: I1125 16:49:37.948245 4802 patch_prober.go:28] interesting pod/downloads-7954f5f757-z8dgw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 25 16:49:37 crc kubenswrapper[4802]: I1125 16:49:37.948800 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-z8dgw" podUID="8a3ebc4f-0491-4dc1-8c5d-adb851b6ef87" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 25 16:49:37 crc kubenswrapper[4802]: I1125 16:49:37.948291 4802 patch_prober.go:28] interesting pod/downloads-7954f5f757-z8dgw container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 25 16:49:37 crc kubenswrapper[4802]: I1125 16:49:37.948918 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-z8dgw" podUID="8a3ebc4f-0491-4dc1-8c5d-adb851b6ef87" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 25 16:49:38 crc kubenswrapper[4802]: I1125 16:49:38.004407 4802 patch_prober.go:28] interesting pod/console-f9d7485db-p7zvz container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 25 16:49:38 crc kubenswrapper[4802]: I1125 16:49:38.004481 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-p7zvz" podUID="df33868c-7e7c-4d67-b478-40ee1d7ef69d" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 25 16:49:40 crc kubenswrapper[4802]: I1125 16:49:40.626232 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs\") pod \"network-metrics-daemon-6ft2z\" (UID: \"cd9aa596-76c2-468c-b732-d85bc1f16070\") " pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:49:40 crc kubenswrapper[4802]: I1125 16:49:40.634026 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cd9aa596-76c2-468c-b732-d85bc1f16070-metrics-certs\") pod \"network-metrics-daemon-6ft2z\" (UID: \"cd9aa596-76c2-468c-b732-d85bc1f16070\") " pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:49:40 crc kubenswrapper[4802]: I1125 16:49:40.852859 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 16:49:40 crc kubenswrapper[4802]: I1125 16:49:40.926054 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6ft2z" Nov 25 16:49:41 crc kubenswrapper[4802]: I1125 16:49:41.032805 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/73a060c4-b1a7-4d9b-875c-e0c426e2fa4f-kubelet-dir\") pod \"73a060c4-b1a7-4d9b-875c-e0c426e2fa4f\" (UID: \"73a060c4-b1a7-4d9b-875c-e0c426e2fa4f\") " Nov 25 16:49:41 crc kubenswrapper[4802]: I1125 16:49:41.032940 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/73a060c4-b1a7-4d9b-875c-e0c426e2fa4f-kube-api-access\") pod \"73a060c4-b1a7-4d9b-875c-e0c426e2fa4f\" (UID: \"73a060c4-b1a7-4d9b-875c-e0c426e2fa4f\") " Nov 25 16:49:41 crc kubenswrapper[4802]: I1125 16:49:41.032980 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73a060c4-b1a7-4d9b-875c-e0c426e2fa4f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "73a060c4-b1a7-4d9b-875c-e0c426e2fa4f" (UID: "73a060c4-b1a7-4d9b-875c-e0c426e2fa4f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:49:41 crc kubenswrapper[4802]: I1125 16:49:41.038622 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73a060c4-b1a7-4d9b-875c-e0c426e2fa4f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "73a060c4-b1a7-4d9b-875c-e0c426e2fa4f" (UID: "73a060c4-b1a7-4d9b-875c-e0c426e2fa4f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:49:41 crc kubenswrapper[4802]: I1125 16:49:41.135031 4802 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/73a060c4-b1a7-4d9b-875c-e0c426e2fa4f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 16:49:41 crc kubenswrapper[4802]: I1125 16:49:41.135193 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/73a060c4-b1a7-4d9b-875c-e0c426e2fa4f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 16:49:41 crc kubenswrapper[4802]: I1125 16:49:41.243599 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"73a060c4-b1a7-4d9b-875c-e0c426e2fa4f","Type":"ContainerDied","Data":"960e8ac4d7ab14413d4cef40cc5a19da7db93d71741190f17d9a114a5788dfdb"} Nov 25 16:49:41 crc kubenswrapper[4802]: I1125 16:49:41.243659 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="960e8ac4d7ab14413d4cef40cc5a19da7db93d71741190f17d9a114a5788dfdb" Nov 25 16:49:41 crc kubenswrapper[4802]: I1125 16:49:41.244147 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 16:49:48 crc kubenswrapper[4802]: I1125 16:49:48.001205 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-z8dgw" Nov 25 16:49:48 crc kubenswrapper[4802]: I1125 16:49:48.008241 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:48 crc kubenswrapper[4802]: I1125 16:49:48.015068 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-p7zvz" Nov 25 16:49:48 crc kubenswrapper[4802]: I1125 16:49:48.058563 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:49:54 crc kubenswrapper[4802]: I1125 16:49:54.248415 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:49:54 crc kubenswrapper[4802]: I1125 16:49:54.249724 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:49:58 crc kubenswrapper[4802]: I1125 16:49:58.562972 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zjrnl" Nov 25 16:50:00 crc kubenswrapper[4802]: E1125 16:50:00.539372 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 25 16:50:00 crc kubenswrapper[4802]: E1125 16:50:00.539573 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qm7mf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-r9nw2_openshift-marketplace(9b19f412-72dd-4fe9-b0f1-3abc306653ae): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 16:50:00 crc kubenswrapper[4802]: E1125 16:50:00.540880 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-r9nw2" podUID="9b19f412-72dd-4fe9-b0f1-3abc306653ae" Nov 25 16:50:03 crc kubenswrapper[4802]: E1125 16:50:03.201716 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-r9nw2" podUID="9b19f412-72dd-4fe9-b0f1-3abc306653ae" Nov 25 16:50:04 crc kubenswrapper[4802]: I1125 16:50:04.993681 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 16:50:08 crc kubenswrapper[4802]: E1125 16:50:08.092752 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 25 16:50:08 crc kubenswrapper[4802]: E1125 16:50:08.093364 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sr8ld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-469c8_openshift-marketplace(8790f862-7ccc-4b14-8c99-1f64349079e0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 16:50:08 crc kubenswrapper[4802]: E1125 16:50:08.094704 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-469c8" podUID="8790f862-7ccc-4b14-8c99-1f64349079e0" Nov 25 16:50:17 crc kubenswrapper[4802]: E1125 16:50:17.535575 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 25 16:50:17 crc kubenswrapper[4802]: E1125 16:50:17.536826 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rxjt5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-jrq5g_openshift-marketplace(ac5f001d-89fc-41e3-8a80-527e81eb55d0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 16:50:17 crc kubenswrapper[4802]: E1125 16:50:17.538181 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-jrq5g" podUID="ac5f001d-89fc-41e3-8a80-527e81eb55d0" Nov 25 16:50:19 crc kubenswrapper[4802]: E1125 16:50:19.277423 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-jrq5g" podUID="ac5f001d-89fc-41e3-8a80-527e81eb55d0" Nov 25 16:50:19 crc kubenswrapper[4802]: W1125 16:50:19.517048 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd9aa596_76c2_468c_b732_d85bc1f16070.slice/crio-1cc8db5f197b56db72f46d4c0d4303e9a3989e528e3ab36149eae3bb9ece404e WatchSource:0}: Error finding container 1cc8db5f197b56db72f46d4c0d4303e9a3989e528e3ab36149eae3bb9ece404e: Status 404 returned error can't find the container with id 1cc8db5f197b56db72f46d4c0d4303e9a3989e528e3ab36149eae3bb9ece404e Nov 25 16:50:19 crc kubenswrapper[4802]: I1125 16:50:19.517235 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-6ft2z"] Nov 25 16:50:20 crc kubenswrapper[4802]: I1125 16:50:20.523803 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" event={"ID":"cd9aa596-76c2-468c-b732-d85bc1f16070","Type":"ContainerStarted","Data":"1cc8db5f197b56db72f46d4c0d4303e9a3989e528e3ab36149eae3bb9ece404e"} Nov 25 16:50:21 crc kubenswrapper[4802]: I1125 16:50:21.532435 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" event={"ID":"cd9aa596-76c2-468c-b732-d85bc1f16070","Type":"ContainerStarted","Data":"01042bbaa020313efe3e10e5fbb51ffd3001d355f19b4768493f4765003a710d"} Nov 25 16:50:21 crc kubenswrapper[4802]: E1125 16:50:21.733563 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 25 16:50:21 crc kubenswrapper[4802]: E1125 16:50:21.733780 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4h6dr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-rtgv8_openshift-marketplace(565aa173-3fac-4f4e-8fae-3fbe88184efb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 16:50:21 crc kubenswrapper[4802]: E1125 16:50:21.734957 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-rtgv8" podUID="565aa173-3fac-4f4e-8fae-3fbe88184efb" Nov 25 16:50:22 crc kubenswrapper[4802]: E1125 16:50:22.546163 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-rtgv8" podUID="565aa173-3fac-4f4e-8fae-3fbe88184efb" Nov 25 16:50:24 crc kubenswrapper[4802]: E1125 16:50:24.166481 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 25 16:50:24 crc kubenswrapper[4802]: E1125 16:50:24.167208 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kbnvk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-gvlq8_openshift-marketplace(ce7b5945-011b-46bb-9af5-96fe9d8e0f89): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 16:50:24 crc kubenswrapper[4802]: E1125 16:50:24.168510 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-gvlq8" podUID="ce7b5945-011b-46bb-9af5-96fe9d8e0f89" Nov 25 16:50:24 crc kubenswrapper[4802]: I1125 16:50:24.249826 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:50:24 crc kubenswrapper[4802]: I1125 16:50:24.250420 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:50:24 crc kubenswrapper[4802]: I1125 16:50:24.250688 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 16:50:24 crc kubenswrapper[4802]: I1125 16:50:24.251996 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef"} pod="openshift-machine-config-operator/machine-config-daemon-h29wc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 16:50:24 crc kubenswrapper[4802]: I1125 16:50:24.252554 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" containerID="cri-o://5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef" gracePeriod=600 Nov 25 16:50:24 crc kubenswrapper[4802]: E1125 16:50:24.375228 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 25 16:50:24 crc kubenswrapper[4802]: E1125 16:50:24.375446 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8dvqh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-nhwgv_openshift-marketplace(078b7cce-5ee4-4c6e-9d9b-5332a6f070ff): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 16:50:24 crc kubenswrapper[4802]: E1125 16:50:24.376752 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-nhwgv" podUID="078b7cce-5ee4-4c6e-9d9b-5332a6f070ff" Nov 25 16:50:24 crc kubenswrapper[4802]: I1125 16:50:24.557018 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-6ft2z" event={"ID":"cd9aa596-76c2-468c-b732-d85bc1f16070","Type":"ContainerStarted","Data":"ffc2bfa305c520127a1cf8173e1011825d9ebc25df11c719f90c50dcdbad5d0c"} Nov 25 16:50:24 crc kubenswrapper[4802]: I1125 16:50:24.600869 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-6ft2z" podStartSLOduration=186.600842931 podStartE2EDuration="3m6.600842931s" podCreationTimestamp="2025-11-25 16:47:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:50:24.577896648 +0000 UTC m=+207.722243834" watchObservedRunningTime="2025-11-25 16:50:24.600842931 +0000 UTC m=+207.745190117" Nov 25 16:50:25 crc kubenswrapper[4802]: I1125 16:50:25.566332 4802 generic.go:334] "Generic (PLEG): container finished" podID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerID="5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef" exitCode=0 Nov 25 16:50:25 crc kubenswrapper[4802]: I1125 16:50:25.566456 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerDied","Data":"5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef"} Nov 25 16:50:36 crc kubenswrapper[4802]: E1125 16:50:36.117145 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 25 16:50:36 crc kubenswrapper[4802]: E1125 16:50:36.118151 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cq2wd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-t56jf_openshift-marketplace(31e04778-8768-4164-a76d-8afc694db820): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 16:50:36 crc kubenswrapper[4802]: E1125 16:50:36.119381 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-t56jf" podUID="31e04778-8768-4164-a76d-8afc694db820" Nov 25 16:50:36 crc kubenswrapper[4802]: E1125 16:50:36.170820 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 25 16:50:36 crc kubenswrapper[4802]: E1125 16:50:36.172470 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4ckk8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-wsq47_openshift-marketplace(55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 16:50:36 crc kubenswrapper[4802]: E1125 16:50:36.174150 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-wsq47" podUID="55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21" Nov 25 16:50:37 crc kubenswrapper[4802]: E1125 16:50:37.784590 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-t56jf" podUID="31e04778-8768-4164-a76d-8afc694db820" Nov 25 16:50:37 crc kubenswrapper[4802]: E1125 16:50:37.784630 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-wsq47" podUID="55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21" Nov 25 16:50:39 crc kubenswrapper[4802]: I1125 16:50:39.653115 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerStarted","Data":"23620d61a8ef1bbe68d551d6cb4fb38ea02979670932fc175c7a86c5b2d8155b"} Nov 25 16:50:40 crc kubenswrapper[4802]: I1125 16:50:40.660964 4802 generic.go:334] "Generic (PLEG): container finished" podID="9b19f412-72dd-4fe9-b0f1-3abc306653ae" containerID="e1767772901a233e33ef62e16d14a3fc767a43eac511a4d01dd4f8c98542cc20" exitCode=0 Nov 25 16:50:40 crc kubenswrapper[4802]: I1125 16:50:40.661017 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9nw2" event={"ID":"9b19f412-72dd-4fe9-b0f1-3abc306653ae","Type":"ContainerDied","Data":"e1767772901a233e33ef62e16d14a3fc767a43eac511a4d01dd4f8c98542cc20"} Nov 25 16:50:40 crc kubenswrapper[4802]: I1125 16:50:40.664307 4802 generic.go:334] "Generic (PLEG): container finished" podID="078b7cce-5ee4-4c6e-9d9b-5332a6f070ff" containerID="63a5989bcd323b5b386514a5d6bb8fb70fa9eddcfb92d71e0f32b30cac00d387" exitCode=0 Nov 25 16:50:40 crc kubenswrapper[4802]: I1125 16:50:40.664377 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhwgv" event={"ID":"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff","Type":"ContainerDied","Data":"63a5989bcd323b5b386514a5d6bb8fb70fa9eddcfb92d71e0f32b30cac00d387"} Nov 25 16:50:40 crc kubenswrapper[4802]: I1125 16:50:40.666853 4802 generic.go:334] "Generic (PLEG): container finished" podID="8790f862-7ccc-4b14-8c99-1f64349079e0" containerID="98949a755bbc45a3e7660dfc427e538d79817ac5678486ce554d0814e3f66253" exitCode=0 Nov 25 16:50:40 crc kubenswrapper[4802]: I1125 16:50:40.666915 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-469c8" event={"ID":"8790f862-7ccc-4b14-8c99-1f64349079e0","Type":"ContainerDied","Data":"98949a755bbc45a3e7660dfc427e538d79817ac5678486ce554d0814e3f66253"} Nov 25 16:50:40 crc kubenswrapper[4802]: I1125 16:50:40.670153 4802 generic.go:334] "Generic (PLEG): container finished" podID="ac5f001d-89fc-41e3-8a80-527e81eb55d0" containerID="c31fd806a60913466379faccf4994f674de54b89d2559fbc7455e3838dac9bba" exitCode=0 Nov 25 16:50:40 crc kubenswrapper[4802]: I1125 16:50:40.670223 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jrq5g" event={"ID":"ac5f001d-89fc-41e3-8a80-527e81eb55d0","Type":"ContainerDied","Data":"c31fd806a60913466379faccf4994f674de54b89d2559fbc7455e3838dac9bba"} Nov 25 16:50:40 crc kubenswrapper[4802]: I1125 16:50:40.676176 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rtgv8" event={"ID":"565aa173-3fac-4f4e-8fae-3fbe88184efb","Type":"ContainerStarted","Data":"7e4823bc4a29a760bd06c65226babca4c32332c8bd477fcf7f85d3406c3e78ee"} Nov 25 16:50:40 crc kubenswrapper[4802]: I1125 16:50:40.684721 4802 generic.go:334] "Generic (PLEG): container finished" podID="ce7b5945-011b-46bb-9af5-96fe9d8e0f89" containerID="c85c15436ffd9b8018da35bea6acde18d6eddbd1d24e2a298ccbbb93556a530b" exitCode=0 Nov 25 16:50:40 crc kubenswrapper[4802]: I1125 16:50:40.684803 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gvlq8" event={"ID":"ce7b5945-011b-46bb-9af5-96fe9d8e0f89","Type":"ContainerDied","Data":"c85c15436ffd9b8018da35bea6acde18d6eddbd1d24e2a298ccbbb93556a530b"} Nov 25 16:50:41 crc kubenswrapper[4802]: I1125 16:50:41.693608 4802 generic.go:334] "Generic (PLEG): container finished" podID="565aa173-3fac-4f4e-8fae-3fbe88184efb" containerID="7e4823bc4a29a760bd06c65226babca4c32332c8bd477fcf7f85d3406c3e78ee" exitCode=0 Nov 25 16:50:41 crc kubenswrapper[4802]: I1125 16:50:41.693712 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rtgv8" event={"ID":"565aa173-3fac-4f4e-8fae-3fbe88184efb","Type":"ContainerDied","Data":"7e4823bc4a29a760bd06c65226babca4c32332c8bd477fcf7f85d3406c3e78ee"} Nov 25 16:50:42 crc kubenswrapper[4802]: I1125 16:50:42.704853 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9nw2" event={"ID":"9b19f412-72dd-4fe9-b0f1-3abc306653ae","Type":"ContainerStarted","Data":"687abc839c294e282339ef0bfeb71caeaa59d38925acff4513fec772540ec13f"} Nov 25 16:50:42 crc kubenswrapper[4802]: I1125 16:50:42.730804 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r9nw2" podStartSLOduration=4.193218586 podStartE2EDuration="1m17.730768836s" podCreationTimestamp="2025-11-25 16:49:25 +0000 UTC" firstStartedPulling="2025-11-25 16:49:27.975392642 +0000 UTC m=+151.119739828" lastFinishedPulling="2025-11-25 16:50:41.512942862 +0000 UTC m=+224.657290078" observedRunningTime="2025-11-25 16:50:42.726092256 +0000 UTC m=+225.870439452" watchObservedRunningTime="2025-11-25 16:50:42.730768836 +0000 UTC m=+225.875116032" Nov 25 16:50:43 crc kubenswrapper[4802]: I1125 16:50:43.713651 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gvlq8" event={"ID":"ce7b5945-011b-46bb-9af5-96fe9d8e0f89","Type":"ContainerStarted","Data":"84b083c8df6233a0d15cffcafac76284b8b8c617bccc7d28a705e5ee0e61a8c9"} Nov 25 16:50:43 crc kubenswrapper[4802]: I1125 16:50:43.733370 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gvlq8" podStartSLOduration=3.984320709 podStartE2EDuration="1m16.733344474s" podCreationTimestamp="2025-11-25 16:49:27 +0000 UTC" firstStartedPulling="2025-11-25 16:49:30.128433418 +0000 UTC m=+153.272780604" lastFinishedPulling="2025-11-25 16:50:42.877457183 +0000 UTC m=+226.021804369" observedRunningTime="2025-11-25 16:50:43.731741069 +0000 UTC m=+226.876088255" watchObservedRunningTime="2025-11-25 16:50:43.733344474 +0000 UTC m=+226.877691660" Nov 25 16:50:44 crc kubenswrapper[4802]: I1125 16:50:44.721655 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jrq5g" event={"ID":"ac5f001d-89fc-41e3-8a80-527e81eb55d0","Type":"ContainerStarted","Data":"2503cf2b8237cdb595282fdc29c696992c724c866215d4c963d2e925f9fb8b75"} Nov 25 16:50:44 crc kubenswrapper[4802]: I1125 16:50:44.735993 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rtgv8" event={"ID":"565aa173-3fac-4f4e-8fae-3fbe88184efb","Type":"ContainerStarted","Data":"ccf1b5358fa3e74a0152ef8ccd66173125f86b34474a6386b043cb45c1f1332a"} Nov 25 16:50:44 crc kubenswrapper[4802]: I1125 16:50:44.743850 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhwgv" event={"ID":"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff","Type":"ContainerStarted","Data":"c0638cfa64aacd76ba25a02b71ef3851eb31b0f0fceda50c3111a9b8ba09b2c8"} Nov 25 16:50:44 crc kubenswrapper[4802]: I1125 16:50:44.780681 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jrq5g" podStartSLOduration=3.97482354 podStartE2EDuration="1m19.780662304s" podCreationTimestamp="2025-11-25 16:49:25 +0000 UTC" firstStartedPulling="2025-11-25 16:49:27.95320249 +0000 UTC m=+151.097549666" lastFinishedPulling="2025-11-25 16:50:43.759041244 +0000 UTC m=+226.903388430" observedRunningTime="2025-11-25 16:50:44.759264565 +0000 UTC m=+227.903611761" watchObservedRunningTime="2025-11-25 16:50:44.780662304 +0000 UTC m=+227.925009490" Nov 25 16:50:44 crc kubenswrapper[4802]: I1125 16:50:44.806236 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nhwgv" podStartSLOduration=3.131332851 podStartE2EDuration="1m18.806212999s" podCreationTimestamp="2025-11-25 16:49:26 +0000 UTC" firstStartedPulling="2025-11-25 16:49:27.988698473 +0000 UTC m=+151.133045659" lastFinishedPulling="2025-11-25 16:50:43.663578621 +0000 UTC m=+226.807925807" observedRunningTime="2025-11-25 16:50:44.781241431 +0000 UTC m=+227.925588647" watchObservedRunningTime="2025-11-25 16:50:44.806212999 +0000 UTC m=+227.950560185" Nov 25 16:50:44 crc kubenswrapper[4802]: I1125 16:50:44.807891 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rtgv8" podStartSLOduration=3.941315727 podStartE2EDuration="1m19.807885727s" podCreationTimestamp="2025-11-25 16:49:25 +0000 UTC" firstStartedPulling="2025-11-25 16:49:27.95982454 +0000 UTC m=+151.104171726" lastFinishedPulling="2025-11-25 16:50:43.82639453 +0000 UTC m=+226.970741726" observedRunningTime="2025-11-25 16:50:44.804212364 +0000 UTC m=+227.948559560" watchObservedRunningTime="2025-11-25 16:50:44.807885727 +0000 UTC m=+227.952232913" Nov 25 16:50:45 crc kubenswrapper[4802]: I1125 16:50:45.753374 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-469c8" event={"ID":"8790f862-7ccc-4b14-8c99-1f64349079e0","Type":"ContainerStarted","Data":"b49b627216d874a034e39e50347a94efad8726d938111a0562a4918720c3093c"} Nov 25 16:50:45 crc kubenswrapper[4802]: I1125 16:50:45.779090 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-469c8" podStartSLOduration=3.171270596 podStartE2EDuration="1m18.779034694s" podCreationTimestamp="2025-11-25 16:49:27 +0000 UTC" firstStartedPulling="2025-11-25 16:49:29.062439774 +0000 UTC m=+152.206786960" lastFinishedPulling="2025-11-25 16:50:44.670203872 +0000 UTC m=+227.814551058" observedRunningTime="2025-11-25 16:50:45.774195169 +0000 UTC m=+228.918542385" watchObservedRunningTime="2025-11-25 16:50:45.779034694 +0000 UTC m=+228.923381890" Nov 25 16:50:45 crc kubenswrapper[4802]: I1125 16:50:45.932318 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r9nw2" Nov 25 16:50:45 crc kubenswrapper[4802]: I1125 16:50:45.932381 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r9nw2" Nov 25 16:50:46 crc kubenswrapper[4802]: I1125 16:50:46.437656 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jrq5g" Nov 25 16:50:46 crc kubenswrapper[4802]: I1125 16:50:46.437744 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jrq5g" Nov 25 16:50:46 crc kubenswrapper[4802]: I1125 16:50:46.438732 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rtgv8" Nov 25 16:50:46 crc kubenswrapper[4802]: I1125 16:50:46.438793 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rtgv8" Nov 25 16:50:46 crc kubenswrapper[4802]: I1125 16:50:46.525286 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nhwgv" Nov 25 16:50:46 crc kubenswrapper[4802]: I1125 16:50:46.525365 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nhwgv" Nov 25 16:50:47 crc kubenswrapper[4802]: I1125 16:50:47.389865 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rtgv8" Nov 25 16:50:47 crc kubenswrapper[4802]: I1125 16:50:47.389975 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nhwgv" Nov 25 16:50:47 crc kubenswrapper[4802]: I1125 16:50:47.390988 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r9nw2" Nov 25 16:50:47 crc kubenswrapper[4802]: I1125 16:50:47.391878 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jrq5g" Nov 25 16:50:47 crc kubenswrapper[4802]: I1125 16:50:47.462793 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r9nw2" Nov 25 16:50:47 crc kubenswrapper[4802]: I1125 16:50:47.896274 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-469c8" Nov 25 16:50:47 crc kubenswrapper[4802]: I1125 16:50:47.897015 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-469c8" Nov 25 16:50:47 crc kubenswrapper[4802]: I1125 16:50:47.964699 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-469c8" Nov 25 16:50:48 crc kubenswrapper[4802]: I1125 16:50:48.623010 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gvlq8" Nov 25 16:50:48 crc kubenswrapper[4802]: I1125 16:50:48.623064 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gvlq8" Nov 25 16:50:48 crc kubenswrapper[4802]: I1125 16:50:48.665802 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gvlq8" Nov 25 16:50:48 crc kubenswrapper[4802]: I1125 16:50:48.817918 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gvlq8" Nov 25 16:50:49 crc kubenswrapper[4802]: I1125 16:50:49.295880 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gvlq8"] Nov 25 16:50:50 crc kubenswrapper[4802]: I1125 16:50:50.788728 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gvlq8" podUID="ce7b5945-011b-46bb-9af5-96fe9d8e0f89" containerName="registry-server" containerID="cri-o://84b083c8df6233a0d15cffcafac76284b8b8c617bccc7d28a705e5ee0e61a8c9" gracePeriod=2 Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.334223 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gvlq8" Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.500189 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbnvk\" (UniqueName: \"kubernetes.io/projected/ce7b5945-011b-46bb-9af5-96fe9d8e0f89-kube-api-access-kbnvk\") pod \"ce7b5945-011b-46bb-9af5-96fe9d8e0f89\" (UID: \"ce7b5945-011b-46bb-9af5-96fe9d8e0f89\") " Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.500285 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce7b5945-011b-46bb-9af5-96fe9d8e0f89-catalog-content\") pod \"ce7b5945-011b-46bb-9af5-96fe9d8e0f89\" (UID: \"ce7b5945-011b-46bb-9af5-96fe9d8e0f89\") " Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.500354 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce7b5945-011b-46bb-9af5-96fe9d8e0f89-utilities\") pod \"ce7b5945-011b-46bb-9af5-96fe9d8e0f89\" (UID: \"ce7b5945-011b-46bb-9af5-96fe9d8e0f89\") " Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.509556 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce7b5945-011b-46bb-9af5-96fe9d8e0f89-utilities" (OuterVolumeSpecName: "utilities") pod "ce7b5945-011b-46bb-9af5-96fe9d8e0f89" (UID: "ce7b5945-011b-46bb-9af5-96fe9d8e0f89"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.537195 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce7b5945-011b-46bb-9af5-96fe9d8e0f89-kube-api-access-kbnvk" (OuterVolumeSpecName: "kube-api-access-kbnvk") pod "ce7b5945-011b-46bb-9af5-96fe9d8e0f89" (UID: "ce7b5945-011b-46bb-9af5-96fe9d8e0f89"). InnerVolumeSpecName "kube-api-access-kbnvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.543435 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce7b5945-011b-46bb-9af5-96fe9d8e0f89-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce7b5945-011b-46bb-9af5-96fe9d8e0f89" (UID: "ce7b5945-011b-46bb-9af5-96fe9d8e0f89"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.603479 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbnvk\" (UniqueName: \"kubernetes.io/projected/ce7b5945-011b-46bb-9af5-96fe9d8e0f89-kube-api-access-kbnvk\") on node \"crc\" DevicePath \"\"" Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.603517 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce7b5945-011b-46bb-9af5-96fe9d8e0f89-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.603528 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce7b5945-011b-46bb-9af5-96fe9d8e0f89-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.796308 4802 generic.go:334] "Generic (PLEG): container finished" podID="ce7b5945-011b-46bb-9af5-96fe9d8e0f89" containerID="84b083c8df6233a0d15cffcafac76284b8b8c617bccc7d28a705e5ee0e61a8c9" exitCode=0 Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.796395 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gvlq8" event={"ID":"ce7b5945-011b-46bb-9af5-96fe9d8e0f89","Type":"ContainerDied","Data":"84b083c8df6233a0d15cffcafac76284b8b8c617bccc7d28a705e5ee0e61a8c9"} Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.796432 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gvlq8" event={"ID":"ce7b5945-011b-46bb-9af5-96fe9d8e0f89","Type":"ContainerDied","Data":"b50e23a0369c8f32298eb164697f2220bc192cdb1613afe62311e1f17197165d"} Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.796458 4802 scope.go:117] "RemoveContainer" containerID="84b083c8df6233a0d15cffcafac76284b8b8c617bccc7d28a705e5ee0e61a8c9" Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.796477 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gvlq8" Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.800343 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsq47" event={"ID":"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21","Type":"ContainerStarted","Data":"2ff38bf38de72c6caeaa5608f69f032884570e0171406fa14892004c20d88b53"} Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.821696 4802 scope.go:117] "RemoveContainer" containerID="c85c15436ffd9b8018da35bea6acde18d6eddbd1d24e2a298ccbbb93556a530b" Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.845615 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gvlq8"] Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.849733 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gvlq8"] Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.866396 4802 scope.go:117] "RemoveContainer" containerID="c018dfd930fd159185fa7f25c5a21a7b4ab38ee055480cfb4428ae5efc61bd97" Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.881942 4802 scope.go:117] "RemoveContainer" containerID="84b083c8df6233a0d15cffcafac76284b8b8c617bccc7d28a705e5ee0e61a8c9" Nov 25 16:50:51 crc kubenswrapper[4802]: E1125 16:50:51.882492 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84b083c8df6233a0d15cffcafac76284b8b8c617bccc7d28a705e5ee0e61a8c9\": container with ID starting with 84b083c8df6233a0d15cffcafac76284b8b8c617bccc7d28a705e5ee0e61a8c9 not found: ID does not exist" containerID="84b083c8df6233a0d15cffcafac76284b8b8c617bccc7d28a705e5ee0e61a8c9" Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.882559 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84b083c8df6233a0d15cffcafac76284b8b8c617bccc7d28a705e5ee0e61a8c9"} err="failed to get container status \"84b083c8df6233a0d15cffcafac76284b8b8c617bccc7d28a705e5ee0e61a8c9\": rpc error: code = NotFound desc = could not find container \"84b083c8df6233a0d15cffcafac76284b8b8c617bccc7d28a705e5ee0e61a8c9\": container with ID starting with 84b083c8df6233a0d15cffcafac76284b8b8c617bccc7d28a705e5ee0e61a8c9 not found: ID does not exist" Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.882605 4802 scope.go:117] "RemoveContainer" containerID="c85c15436ffd9b8018da35bea6acde18d6eddbd1d24e2a298ccbbb93556a530b" Nov 25 16:50:51 crc kubenswrapper[4802]: E1125 16:50:51.882929 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c85c15436ffd9b8018da35bea6acde18d6eddbd1d24e2a298ccbbb93556a530b\": container with ID starting with c85c15436ffd9b8018da35bea6acde18d6eddbd1d24e2a298ccbbb93556a530b not found: ID does not exist" containerID="c85c15436ffd9b8018da35bea6acde18d6eddbd1d24e2a298ccbbb93556a530b" Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.882970 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c85c15436ffd9b8018da35bea6acde18d6eddbd1d24e2a298ccbbb93556a530b"} err="failed to get container status \"c85c15436ffd9b8018da35bea6acde18d6eddbd1d24e2a298ccbbb93556a530b\": rpc error: code = NotFound desc = could not find container \"c85c15436ffd9b8018da35bea6acde18d6eddbd1d24e2a298ccbbb93556a530b\": container with ID starting with c85c15436ffd9b8018da35bea6acde18d6eddbd1d24e2a298ccbbb93556a530b not found: ID does not exist" Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.882998 4802 scope.go:117] "RemoveContainer" containerID="c018dfd930fd159185fa7f25c5a21a7b4ab38ee055480cfb4428ae5efc61bd97" Nov 25 16:50:51 crc kubenswrapper[4802]: E1125 16:50:51.883429 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c018dfd930fd159185fa7f25c5a21a7b4ab38ee055480cfb4428ae5efc61bd97\": container with ID starting with c018dfd930fd159185fa7f25c5a21a7b4ab38ee055480cfb4428ae5efc61bd97 not found: ID does not exist" containerID="c018dfd930fd159185fa7f25c5a21a7b4ab38ee055480cfb4428ae5efc61bd97" Nov 25 16:50:51 crc kubenswrapper[4802]: I1125 16:50:51.883473 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c018dfd930fd159185fa7f25c5a21a7b4ab38ee055480cfb4428ae5efc61bd97"} err="failed to get container status \"c018dfd930fd159185fa7f25c5a21a7b4ab38ee055480cfb4428ae5efc61bd97\": rpc error: code = NotFound desc = could not find container \"c018dfd930fd159185fa7f25c5a21a7b4ab38ee055480cfb4428ae5efc61bd97\": container with ID starting with c018dfd930fd159185fa7f25c5a21a7b4ab38ee055480cfb4428ae5efc61bd97 not found: ID does not exist" Nov 25 16:50:52 crc kubenswrapper[4802]: I1125 16:50:52.810236 4802 generic.go:334] "Generic (PLEG): container finished" podID="55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21" containerID="2ff38bf38de72c6caeaa5608f69f032884570e0171406fa14892004c20d88b53" exitCode=0 Nov 25 16:50:52 crc kubenswrapper[4802]: I1125 16:50:52.810319 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsq47" event={"ID":"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21","Type":"ContainerDied","Data":"2ff38bf38de72c6caeaa5608f69f032884570e0171406fa14892004c20d88b53"} Nov 25 16:50:52 crc kubenswrapper[4802]: I1125 16:50:52.815166 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t56jf" event={"ID":"31e04778-8768-4164-a76d-8afc694db820","Type":"ContainerStarted","Data":"09aeee779caabdebfb77bbefa06ab536e2dadc9f5e00ebcfdf6a7f2be9b738bb"} Nov 25 16:50:53 crc kubenswrapper[4802]: I1125 16:50:53.513461 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce7b5945-011b-46bb-9af5-96fe9d8e0f89" path="/var/lib/kubelet/pods/ce7b5945-011b-46bb-9af5-96fe9d8e0f89/volumes" Nov 25 16:50:53 crc kubenswrapper[4802]: I1125 16:50:53.835687 4802 generic.go:334] "Generic (PLEG): container finished" podID="31e04778-8768-4164-a76d-8afc694db820" containerID="09aeee779caabdebfb77bbefa06ab536e2dadc9f5e00ebcfdf6a7f2be9b738bb" exitCode=0 Nov 25 16:50:53 crc kubenswrapper[4802]: I1125 16:50:53.835796 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t56jf" event={"ID":"31e04778-8768-4164-a76d-8afc694db820","Type":"ContainerDied","Data":"09aeee779caabdebfb77bbefa06ab536e2dadc9f5e00ebcfdf6a7f2be9b738bb"} Nov 25 16:50:56 crc kubenswrapper[4802]: I1125 16:50:56.512779 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rtgv8" Nov 25 16:50:56 crc kubenswrapper[4802]: I1125 16:50:56.518695 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jrq5g" Nov 25 16:50:56 crc kubenswrapper[4802]: I1125 16:50:56.568270 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nhwgv" Nov 25 16:50:56 crc kubenswrapper[4802]: I1125 16:50:56.858321 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsq47" event={"ID":"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21","Type":"ContainerStarted","Data":"6feb3e15a0c77aa78855831a496db8635c09a073f5e22b439fc5a8d6bd0fafcb"} Nov 25 16:50:57 crc kubenswrapper[4802]: I1125 16:50:57.513775 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jrq5g"] Nov 25 16:50:57 crc kubenswrapper[4802]: I1125 16:50:57.514043 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jrq5g" podUID="ac5f001d-89fc-41e3-8a80-527e81eb55d0" containerName="registry-server" containerID="cri-o://2503cf2b8237cdb595282fdc29c696992c724c866215d4c963d2e925f9fb8b75" gracePeriod=2 Nov 25 16:50:57 crc kubenswrapper[4802]: I1125 16:50:57.889089 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wsq47" podStartSLOduration=5.149838138 podStartE2EDuration="1m29.889067281s" podCreationTimestamp="2025-11-25 16:49:28 +0000 UTC" firstStartedPulling="2025-11-25 16:49:31.154995512 +0000 UTC m=+154.299342698" lastFinishedPulling="2025-11-25 16:50:55.894224655 +0000 UTC m=+239.038571841" observedRunningTime="2025-11-25 16:50:57.886672245 +0000 UTC m=+241.031019451" watchObservedRunningTime="2025-11-25 16:50:57.889067281 +0000 UTC m=+241.033414467" Nov 25 16:50:57 crc kubenswrapper[4802]: I1125 16:50:57.949076 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-469c8" Nov 25 16:50:58 crc kubenswrapper[4802]: I1125 16:50:58.874226 4802 generic.go:334] "Generic (PLEG): container finished" podID="ac5f001d-89fc-41e3-8a80-527e81eb55d0" containerID="2503cf2b8237cdb595282fdc29c696992c724c866215d4c963d2e925f9fb8b75" exitCode=0 Nov 25 16:50:58 crc kubenswrapper[4802]: I1125 16:50:58.874317 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jrq5g" event={"ID":"ac5f001d-89fc-41e3-8a80-527e81eb55d0","Type":"ContainerDied","Data":"2503cf2b8237cdb595282fdc29c696992c724c866215d4c963d2e925f9fb8b75"} Nov 25 16:50:58 crc kubenswrapper[4802]: I1125 16:50:58.898088 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nhwgv"] Nov 25 16:50:58 crc kubenswrapper[4802]: I1125 16:50:58.898693 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nhwgv" podUID="078b7cce-5ee4-4c6e-9d9b-5332a6f070ff" containerName="registry-server" containerID="cri-o://c0638cfa64aacd76ba25a02b71ef3851eb31b0f0fceda50c3111a9b8ba09b2c8" gracePeriod=2 Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.306076 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wsq47" Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.306147 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wsq47" Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.817438 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jrq5g" Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.886245 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jrq5g" event={"ID":"ac5f001d-89fc-41e3-8a80-527e81eb55d0","Type":"ContainerDied","Data":"d3addd95ffa846a7ed83c9b4e3f86402119f755d124ab1f869872b4272e9896e"} Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.886329 4802 scope.go:117] "RemoveContainer" containerID="2503cf2b8237cdb595282fdc29c696992c724c866215d4c963d2e925f9fb8b75" Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.886599 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jrq5g" Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.890204 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t56jf" event={"ID":"31e04778-8768-4164-a76d-8afc694db820","Type":"ContainerStarted","Data":"385a0e3ab0a3bc0837fd95ec6cadb29fabb36c08ac9be608bebaf79f8f228046"} Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.896247 4802 generic.go:334] "Generic (PLEG): container finished" podID="078b7cce-5ee4-4c6e-9d9b-5332a6f070ff" containerID="c0638cfa64aacd76ba25a02b71ef3851eb31b0f0fceda50c3111a9b8ba09b2c8" exitCode=0 Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.896321 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhwgv" event={"ID":"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff","Type":"ContainerDied","Data":"c0638cfa64aacd76ba25a02b71ef3851eb31b0f0fceda50c3111a9b8ba09b2c8"} Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.902804 4802 scope.go:117] "RemoveContainer" containerID="c31fd806a60913466379faccf4994f674de54b89d2559fbc7455e3838dac9bba" Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.915460 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t56jf" podStartSLOduration=2.465830994 podStartE2EDuration="1m30.915434581s" podCreationTimestamp="2025-11-25 16:49:29 +0000 UTC" firstStartedPulling="2025-11-25 16:49:31.155029633 +0000 UTC m=+154.299376819" lastFinishedPulling="2025-11-25 16:50:59.60463322 +0000 UTC m=+242.748980406" observedRunningTime="2025-11-25 16:50:59.913567708 +0000 UTC m=+243.057914894" watchObservedRunningTime="2025-11-25 16:50:59.915434581 +0000 UTC m=+243.059781767" Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.926963 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac5f001d-89fc-41e3-8a80-527e81eb55d0-catalog-content\") pod \"ac5f001d-89fc-41e3-8a80-527e81eb55d0\" (UID: \"ac5f001d-89fc-41e3-8a80-527e81eb55d0\") " Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.927004 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxjt5\" (UniqueName: \"kubernetes.io/projected/ac5f001d-89fc-41e3-8a80-527e81eb55d0-kube-api-access-rxjt5\") pod \"ac5f001d-89fc-41e3-8a80-527e81eb55d0\" (UID: \"ac5f001d-89fc-41e3-8a80-527e81eb55d0\") " Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.927078 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac5f001d-89fc-41e3-8a80-527e81eb55d0-utilities\") pod \"ac5f001d-89fc-41e3-8a80-527e81eb55d0\" (UID: \"ac5f001d-89fc-41e3-8a80-527e81eb55d0\") " Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.928170 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac5f001d-89fc-41e3-8a80-527e81eb55d0-utilities" (OuterVolumeSpecName: "utilities") pod "ac5f001d-89fc-41e3-8a80-527e81eb55d0" (UID: "ac5f001d-89fc-41e3-8a80-527e81eb55d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.928346 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac5f001d-89fc-41e3-8a80-527e81eb55d0-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.929456 4802 scope.go:117] "RemoveContainer" containerID="14ea03ef1504d38cc6ea45818f23bd0acf170243150a6f8f5c15f6b3bbe81f55" Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.942395 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac5f001d-89fc-41e3-8a80-527e81eb55d0-kube-api-access-rxjt5" (OuterVolumeSpecName: "kube-api-access-rxjt5") pod "ac5f001d-89fc-41e3-8a80-527e81eb55d0" (UID: "ac5f001d-89fc-41e3-8a80-527e81eb55d0"). InnerVolumeSpecName "kube-api-access-rxjt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:50:59 crc kubenswrapper[4802]: I1125 16:50:59.977106 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac5f001d-89fc-41e3-8a80-527e81eb55d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac5f001d-89fc-41e3-8a80-527e81eb55d0" (UID: "ac5f001d-89fc-41e3-8a80-527e81eb55d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.042992 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac5f001d-89fc-41e3-8a80-527e81eb55d0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.043041 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxjt5\" (UniqueName: \"kubernetes.io/projected/ac5f001d-89fc-41e3-8a80-527e81eb55d0-kube-api-access-rxjt5\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.221585 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jrq5g"] Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.226408 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jrq5g"] Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.348832 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wsq47" podUID="55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21" containerName="registry-server" probeResult="failure" output=< Nov 25 16:51:00 crc kubenswrapper[4802]: timeout: failed to connect service ":50051" within 1s Nov 25 16:51:00 crc kubenswrapper[4802]: > Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.467825 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nhwgv" Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.655864 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff-utilities\") pod \"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff\" (UID: \"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff\") " Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.656493 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff-catalog-content\") pod \"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff\" (UID: \"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff\") " Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.656734 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dvqh\" (UniqueName: \"kubernetes.io/projected/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff-kube-api-access-8dvqh\") pod \"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff\" (UID: \"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff\") " Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.656802 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff-utilities" (OuterVolumeSpecName: "utilities") pod "078b7cce-5ee4-4c6e-9d9b-5332a6f070ff" (UID: "078b7cce-5ee4-4c6e-9d9b-5332a6f070ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.657354 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.661837 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff-kube-api-access-8dvqh" (OuterVolumeSpecName: "kube-api-access-8dvqh") pod "078b7cce-5ee4-4c6e-9d9b-5332a6f070ff" (UID: "078b7cce-5ee4-4c6e-9d9b-5332a6f070ff"). InnerVolumeSpecName "kube-api-access-8dvqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.711811 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "078b7cce-5ee4-4c6e-9d9b-5332a6f070ff" (UID: "078b7cce-5ee4-4c6e-9d9b-5332a6f070ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.759094 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dvqh\" (UniqueName: \"kubernetes.io/projected/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff-kube-api-access-8dvqh\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.759160 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.912901 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhwgv" event={"ID":"078b7cce-5ee4-4c6e-9d9b-5332a6f070ff","Type":"ContainerDied","Data":"36dfce8c051a60cff6c7ef28623eff7b45d14383b43e4820f4ae1adec82a8227"} Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.912986 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nhwgv" Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.913013 4802 scope.go:117] "RemoveContainer" containerID="c0638cfa64aacd76ba25a02b71ef3851eb31b0f0fceda50c3111a9b8ba09b2c8" Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.952640 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nhwgv"] Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.953104 4802 scope.go:117] "RemoveContainer" containerID="63a5989bcd323b5b386514a5d6bb8fb70fa9eddcfb92d71e0f32b30cac00d387" Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.958249 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nhwgv"] Nov 25 16:51:00 crc kubenswrapper[4802]: I1125 16:51:00.975427 4802 scope.go:117] "RemoveContainer" containerID="750e351dc53ca74511b6d49c9abfc2fbfb9853bb6975c10e76703d7a1fc2d106" Nov 25 16:51:01 crc kubenswrapper[4802]: I1125 16:51:01.511560 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="078b7cce-5ee4-4c6e-9d9b-5332a6f070ff" path="/var/lib/kubelet/pods/078b7cce-5ee4-4c6e-9d9b-5332a6f070ff/volumes" Nov 25 16:51:01 crc kubenswrapper[4802]: I1125 16:51:01.512323 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac5f001d-89fc-41e3-8a80-527e81eb55d0" path="/var/lib/kubelet/pods/ac5f001d-89fc-41e3-8a80-527e81eb55d0/volumes" Nov 25 16:51:09 crc kubenswrapper[4802]: I1125 16:51:09.345781 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wsq47" Nov 25 16:51:09 crc kubenswrapper[4802]: I1125 16:51:09.399392 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wsq47" Nov 25 16:51:09 crc kubenswrapper[4802]: I1125 16:51:09.725426 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t56jf" Nov 25 16:51:09 crc kubenswrapper[4802]: I1125 16:51:09.725985 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t56jf" Nov 25 16:51:09 crc kubenswrapper[4802]: I1125 16:51:09.770639 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t56jf" Nov 25 16:51:10 crc kubenswrapper[4802]: I1125 16:51:10.015978 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t56jf" Nov 25 16:51:10 crc kubenswrapper[4802]: I1125 16:51:10.581774 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t56jf"] Nov 25 16:51:11 crc kubenswrapper[4802]: I1125 16:51:11.986447 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t56jf" podUID="31e04778-8768-4164-a76d-8afc694db820" containerName="registry-server" containerID="cri-o://385a0e3ab0a3bc0837fd95ec6cadb29fabb36c08ac9be608bebaf79f8f228046" gracePeriod=2 Nov 25 16:51:12 crc kubenswrapper[4802]: I1125 16:51:12.339790 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t56jf" Nov 25 16:51:12 crc kubenswrapper[4802]: I1125 16:51:12.523686 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31e04778-8768-4164-a76d-8afc694db820-catalog-content\") pod \"31e04778-8768-4164-a76d-8afc694db820\" (UID: \"31e04778-8768-4164-a76d-8afc694db820\") " Nov 25 16:51:12 crc kubenswrapper[4802]: I1125 16:51:12.524381 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31e04778-8768-4164-a76d-8afc694db820-utilities\") pod \"31e04778-8768-4164-a76d-8afc694db820\" (UID: \"31e04778-8768-4164-a76d-8afc694db820\") " Nov 25 16:51:12 crc kubenswrapper[4802]: I1125 16:51:12.524520 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cq2wd\" (UniqueName: \"kubernetes.io/projected/31e04778-8768-4164-a76d-8afc694db820-kube-api-access-cq2wd\") pod \"31e04778-8768-4164-a76d-8afc694db820\" (UID: \"31e04778-8768-4164-a76d-8afc694db820\") " Nov 25 16:51:12 crc kubenswrapper[4802]: I1125 16:51:12.526346 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31e04778-8768-4164-a76d-8afc694db820-utilities" (OuterVolumeSpecName: "utilities") pod "31e04778-8768-4164-a76d-8afc694db820" (UID: "31e04778-8768-4164-a76d-8afc694db820"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:51:12 crc kubenswrapper[4802]: I1125 16:51:12.531373 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31e04778-8768-4164-a76d-8afc694db820-kube-api-access-cq2wd" (OuterVolumeSpecName: "kube-api-access-cq2wd") pod "31e04778-8768-4164-a76d-8afc694db820" (UID: "31e04778-8768-4164-a76d-8afc694db820"). InnerVolumeSpecName "kube-api-access-cq2wd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:51:12 crc kubenswrapper[4802]: I1125 16:51:12.614440 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31e04778-8768-4164-a76d-8afc694db820-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "31e04778-8768-4164-a76d-8afc694db820" (UID: "31e04778-8768-4164-a76d-8afc694db820"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:51:12 crc kubenswrapper[4802]: I1125 16:51:12.626332 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31e04778-8768-4164-a76d-8afc694db820-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:12 crc kubenswrapper[4802]: I1125 16:51:12.626667 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cq2wd\" (UniqueName: \"kubernetes.io/projected/31e04778-8768-4164-a76d-8afc694db820-kube-api-access-cq2wd\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:12 crc kubenswrapper[4802]: I1125 16:51:12.626771 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31e04778-8768-4164-a76d-8afc694db820-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:12 crc kubenswrapper[4802]: I1125 16:51:12.993830 4802 generic.go:334] "Generic (PLEG): container finished" podID="31e04778-8768-4164-a76d-8afc694db820" containerID="385a0e3ab0a3bc0837fd95ec6cadb29fabb36c08ac9be608bebaf79f8f228046" exitCode=0 Nov 25 16:51:12 crc kubenswrapper[4802]: I1125 16:51:12.993898 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t56jf" event={"ID":"31e04778-8768-4164-a76d-8afc694db820","Type":"ContainerDied","Data":"385a0e3ab0a3bc0837fd95ec6cadb29fabb36c08ac9be608bebaf79f8f228046"} Nov 25 16:51:12 crc kubenswrapper[4802]: I1125 16:51:12.993948 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t56jf" event={"ID":"31e04778-8768-4164-a76d-8afc694db820","Type":"ContainerDied","Data":"25a696ad4e97c044c3dd6ee921273d09badc0af503405e5a6f92b49217e0e5d3"} Nov 25 16:51:12 crc kubenswrapper[4802]: I1125 16:51:12.993968 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t56jf" Nov 25 16:51:12 crc kubenswrapper[4802]: I1125 16:51:12.993974 4802 scope.go:117] "RemoveContainer" containerID="385a0e3ab0a3bc0837fd95ec6cadb29fabb36c08ac9be608bebaf79f8f228046" Nov 25 16:51:13 crc kubenswrapper[4802]: I1125 16:51:13.014157 4802 scope.go:117] "RemoveContainer" containerID="09aeee779caabdebfb77bbefa06ab536e2dadc9f5e00ebcfdf6a7f2be9b738bb" Nov 25 16:51:13 crc kubenswrapper[4802]: I1125 16:51:13.025518 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t56jf"] Nov 25 16:51:13 crc kubenswrapper[4802]: I1125 16:51:13.032327 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t56jf"] Nov 25 16:51:13 crc kubenswrapper[4802]: I1125 16:51:13.058689 4802 scope.go:117] "RemoveContainer" containerID="bfa5d983d60beb0de2d7523d8ac39782d9237fdee78e30471514de377d6e5505" Nov 25 16:51:13 crc kubenswrapper[4802]: I1125 16:51:13.076014 4802 scope.go:117] "RemoveContainer" containerID="385a0e3ab0a3bc0837fd95ec6cadb29fabb36c08ac9be608bebaf79f8f228046" Nov 25 16:51:13 crc kubenswrapper[4802]: E1125 16:51:13.076453 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"385a0e3ab0a3bc0837fd95ec6cadb29fabb36c08ac9be608bebaf79f8f228046\": container with ID starting with 385a0e3ab0a3bc0837fd95ec6cadb29fabb36c08ac9be608bebaf79f8f228046 not found: ID does not exist" containerID="385a0e3ab0a3bc0837fd95ec6cadb29fabb36c08ac9be608bebaf79f8f228046" Nov 25 16:51:13 crc kubenswrapper[4802]: I1125 16:51:13.076572 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"385a0e3ab0a3bc0837fd95ec6cadb29fabb36c08ac9be608bebaf79f8f228046"} err="failed to get container status \"385a0e3ab0a3bc0837fd95ec6cadb29fabb36c08ac9be608bebaf79f8f228046\": rpc error: code = NotFound desc = could not find container \"385a0e3ab0a3bc0837fd95ec6cadb29fabb36c08ac9be608bebaf79f8f228046\": container with ID starting with 385a0e3ab0a3bc0837fd95ec6cadb29fabb36c08ac9be608bebaf79f8f228046 not found: ID does not exist" Nov 25 16:51:13 crc kubenswrapper[4802]: I1125 16:51:13.076639 4802 scope.go:117] "RemoveContainer" containerID="09aeee779caabdebfb77bbefa06ab536e2dadc9f5e00ebcfdf6a7f2be9b738bb" Nov 25 16:51:13 crc kubenswrapper[4802]: E1125 16:51:13.077088 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09aeee779caabdebfb77bbefa06ab536e2dadc9f5e00ebcfdf6a7f2be9b738bb\": container with ID starting with 09aeee779caabdebfb77bbefa06ab536e2dadc9f5e00ebcfdf6a7f2be9b738bb not found: ID does not exist" containerID="09aeee779caabdebfb77bbefa06ab536e2dadc9f5e00ebcfdf6a7f2be9b738bb" Nov 25 16:51:13 crc kubenswrapper[4802]: I1125 16:51:13.077167 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09aeee779caabdebfb77bbefa06ab536e2dadc9f5e00ebcfdf6a7f2be9b738bb"} err="failed to get container status \"09aeee779caabdebfb77bbefa06ab536e2dadc9f5e00ebcfdf6a7f2be9b738bb\": rpc error: code = NotFound desc = could not find container \"09aeee779caabdebfb77bbefa06ab536e2dadc9f5e00ebcfdf6a7f2be9b738bb\": container with ID starting with 09aeee779caabdebfb77bbefa06ab536e2dadc9f5e00ebcfdf6a7f2be9b738bb not found: ID does not exist" Nov 25 16:51:13 crc kubenswrapper[4802]: I1125 16:51:13.077211 4802 scope.go:117] "RemoveContainer" containerID="bfa5d983d60beb0de2d7523d8ac39782d9237fdee78e30471514de377d6e5505" Nov 25 16:51:13 crc kubenswrapper[4802]: E1125 16:51:13.077613 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfa5d983d60beb0de2d7523d8ac39782d9237fdee78e30471514de377d6e5505\": container with ID starting with bfa5d983d60beb0de2d7523d8ac39782d9237fdee78e30471514de377d6e5505 not found: ID does not exist" containerID="bfa5d983d60beb0de2d7523d8ac39782d9237fdee78e30471514de377d6e5505" Nov 25 16:51:13 crc kubenswrapper[4802]: I1125 16:51:13.077651 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfa5d983d60beb0de2d7523d8ac39782d9237fdee78e30471514de377d6e5505"} err="failed to get container status \"bfa5d983d60beb0de2d7523d8ac39782d9237fdee78e30471514de377d6e5505\": rpc error: code = NotFound desc = could not find container \"bfa5d983d60beb0de2d7523d8ac39782d9237fdee78e30471514de377d6e5505\": container with ID starting with bfa5d983d60beb0de2d7523d8ac39782d9237fdee78e30471514de377d6e5505 not found: ID does not exist" Nov 25 16:51:13 crc kubenswrapper[4802]: I1125 16:51:13.513290 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31e04778-8768-4164-a76d-8afc694db820" path="/var/lib/kubelet/pods/31e04778-8768-4164-a76d-8afc694db820/volumes" Nov 25 16:51:21 crc kubenswrapper[4802]: I1125 16:51:21.746722 4802 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-smpjk container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 16:51:21 crc kubenswrapper[4802]: I1125 16:51:21.746756 4802 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-smpjk container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 16:51:21 crc kubenswrapper[4802]: I1125 16:51:21.747793 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" podUID="b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 16:51:21 crc kubenswrapper[4802]: I1125 16:51:21.747898 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-smpjk" podUID="b43f0db4-5ef1-48fc-8ffe-1d5348a2c7c0" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 16:51:28 crc kubenswrapper[4802]: I1125 16:51:28.813018 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sccr2"] Nov 25 16:51:53 crc kubenswrapper[4802]: I1125 16:51:53.850207 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" podUID="63972ba9-04aa-42e8-a91c-f27796ab40f4" containerName="oauth-openshift" containerID="cri-o://bdad6a935ec9721095481ee741472175ca79bc15dff9508448f5f8185ea781be" gracePeriod=15 Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.221824 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.255973 4802 generic.go:334] "Generic (PLEG): container finished" podID="63972ba9-04aa-42e8-a91c-f27796ab40f4" containerID="bdad6a935ec9721095481ee741472175ca79bc15dff9508448f5f8185ea781be" exitCode=0 Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.256047 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" event={"ID":"63972ba9-04aa-42e8-a91c-f27796ab40f4","Type":"ContainerDied","Data":"bdad6a935ec9721095481ee741472175ca79bc15dff9508448f5f8185ea781be"} Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.256101 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" event={"ID":"63972ba9-04aa-42e8-a91c-f27796ab40f4","Type":"ContainerDied","Data":"4f99eed08d483d1ec4ec5f15763112b9d49f1c189d29a229550e517e23341c16"} Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.256163 4802 scope.go:117] "RemoveContainer" containerID="bdad6a935ec9721095481ee741472175ca79bc15dff9508448f5f8185ea781be" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.256620 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sccr2" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.262400 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95"] Nov 25 16:51:54 crc kubenswrapper[4802]: E1125 16:51:54.262660 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac5f001d-89fc-41e3-8a80-527e81eb55d0" containerName="registry-server" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.262685 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac5f001d-89fc-41e3-8a80-527e81eb55d0" containerName="registry-server" Nov 25 16:51:54 crc kubenswrapper[4802]: E1125 16:51:54.262703 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce7b5945-011b-46bb-9af5-96fe9d8e0f89" containerName="registry-server" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.262714 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce7b5945-011b-46bb-9af5-96fe9d8e0f89" containerName="registry-server" Nov 25 16:51:54 crc kubenswrapper[4802]: E1125 16:51:54.262725 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="078b7cce-5ee4-4c6e-9d9b-5332a6f070ff" containerName="extract-content" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.262733 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="078b7cce-5ee4-4c6e-9d9b-5332a6f070ff" containerName="extract-content" Nov 25 16:51:54 crc kubenswrapper[4802]: E1125 16:51:54.262747 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="078b7cce-5ee4-4c6e-9d9b-5332a6f070ff" containerName="extract-utilities" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.262755 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="078b7cce-5ee4-4c6e-9d9b-5332a6f070ff" containerName="extract-utilities" Nov 25 16:51:54 crc kubenswrapper[4802]: E1125 16:51:54.262765 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31e04778-8768-4164-a76d-8afc694db820" containerName="extract-content" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.262772 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="31e04778-8768-4164-a76d-8afc694db820" containerName="extract-content" Nov 25 16:51:54 crc kubenswrapper[4802]: E1125 16:51:54.262784 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31e04778-8768-4164-a76d-8afc694db820" containerName="registry-server" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.262791 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="31e04778-8768-4164-a76d-8afc694db820" containerName="registry-server" Nov 25 16:51:54 crc kubenswrapper[4802]: E1125 16:51:54.262805 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac5f001d-89fc-41e3-8a80-527e81eb55d0" containerName="extract-content" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.262815 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac5f001d-89fc-41e3-8a80-527e81eb55d0" containerName="extract-content" Nov 25 16:51:54 crc kubenswrapper[4802]: E1125 16:51:54.262824 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac5f001d-89fc-41e3-8a80-527e81eb55d0" containerName="extract-utilities" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.262834 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac5f001d-89fc-41e3-8a80-527e81eb55d0" containerName="extract-utilities" Nov 25 16:51:54 crc kubenswrapper[4802]: E1125 16:51:54.262845 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31e04778-8768-4164-a76d-8afc694db820" containerName="extract-utilities" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.262855 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="31e04778-8768-4164-a76d-8afc694db820" containerName="extract-utilities" Nov 25 16:51:54 crc kubenswrapper[4802]: E1125 16:51:54.262864 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce7b5945-011b-46bb-9af5-96fe9d8e0f89" containerName="extract-content" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.262872 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce7b5945-011b-46bb-9af5-96fe9d8e0f89" containerName="extract-content" Nov 25 16:51:54 crc kubenswrapper[4802]: E1125 16:51:54.262887 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63972ba9-04aa-42e8-a91c-f27796ab40f4" containerName="oauth-openshift" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.262895 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="63972ba9-04aa-42e8-a91c-f27796ab40f4" containerName="oauth-openshift" Nov 25 16:51:54 crc kubenswrapper[4802]: E1125 16:51:54.262905 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce7b5945-011b-46bb-9af5-96fe9d8e0f89" containerName="extract-utilities" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.262913 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce7b5945-011b-46bb-9af5-96fe9d8e0f89" containerName="extract-utilities" Nov 25 16:51:54 crc kubenswrapper[4802]: E1125 16:51:54.262921 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="078b7cce-5ee4-4c6e-9d9b-5332a6f070ff" containerName="registry-server" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.262929 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="078b7cce-5ee4-4c6e-9d9b-5332a6f070ff" containerName="registry-server" Nov 25 16:51:54 crc kubenswrapper[4802]: E1125 16:51:54.262940 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73a060c4-b1a7-4d9b-875c-e0c426e2fa4f" containerName="pruner" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.262950 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="73a060c4-b1a7-4d9b-875c-e0c426e2fa4f" containerName="pruner" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.263061 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="31e04778-8768-4164-a76d-8afc694db820" containerName="registry-server" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.263073 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="078b7cce-5ee4-4c6e-9d9b-5332a6f070ff" containerName="registry-server" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.263085 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="73a060c4-b1a7-4d9b-875c-e0c426e2fa4f" containerName="pruner" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.263096 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac5f001d-89fc-41e3-8a80-527e81eb55d0" containerName="registry-server" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.263107 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce7b5945-011b-46bb-9af5-96fe9d8e0f89" containerName="registry-server" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.263117 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="63972ba9-04aa-42e8-a91c-f27796ab40f4" containerName="oauth-openshift" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.265980 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.283312 4802 scope.go:117] "RemoveContainer" containerID="bdad6a935ec9721095481ee741472175ca79bc15dff9508448f5f8185ea781be" Nov 25 16:51:54 crc kubenswrapper[4802]: E1125 16:51:54.284361 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdad6a935ec9721095481ee741472175ca79bc15dff9508448f5f8185ea781be\": container with ID starting with bdad6a935ec9721095481ee741472175ca79bc15dff9508448f5f8185ea781be not found: ID does not exist" containerID="bdad6a935ec9721095481ee741472175ca79bc15dff9508448f5f8185ea781be" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.284434 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdad6a935ec9721095481ee741472175ca79bc15dff9508448f5f8185ea781be"} err="failed to get container status \"bdad6a935ec9721095481ee741472175ca79bc15dff9508448f5f8185ea781be\": rpc error: code = NotFound desc = could not find container \"bdad6a935ec9721095481ee741472175ca79bc15dff9508448f5f8185ea781be\": container with ID starting with bdad6a935ec9721095481ee741472175ca79bc15dff9508448f5f8185ea781be not found: ID does not exist" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.291639 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95"] Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.319167 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.319237 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.319261 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-service-ca\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.319277 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-router-certs\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.319296 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c6df39a7-5145-4f72-bd5e-2bef2077498c-audit-dir\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.319322 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-user-template-error\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.319342 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-user-template-login\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.319365 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.319381 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.319396 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.319737 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkkt6\" (UniqueName: \"kubernetes.io/projected/c6df39a7-5145-4f72-bd5e-2bef2077498c-kube-api-access-rkkt6\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.319795 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-session\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.320174 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.320357 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c6df39a7-5145-4f72-bd5e-2bef2077498c-audit-policies\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.421987 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-cliconfig\") pod \"63972ba9-04aa-42e8-a91c-f27796ab40f4\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.422117 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-template-error\") pod \"63972ba9-04aa-42e8-a91c-f27796ab40f4\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.422214 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-trusted-ca-bundle\") pod \"63972ba9-04aa-42e8-a91c-f27796ab40f4\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.422276 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/63972ba9-04aa-42e8-a91c-f27796ab40f4-audit-dir\") pod \"63972ba9-04aa-42e8-a91c-f27796ab40f4\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.422365 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-template-login\") pod \"63972ba9-04aa-42e8-a91c-f27796ab40f4\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.422434 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-service-ca\") pod \"63972ba9-04aa-42e8-a91c-f27796ab40f4\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.422498 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-serving-cert\") pod \"63972ba9-04aa-42e8-a91c-f27796ab40f4\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.422557 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-router-certs\") pod \"63972ba9-04aa-42e8-a91c-f27796ab40f4\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.422537 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63972ba9-04aa-42e8-a91c-f27796ab40f4-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "63972ba9-04aa-42e8-a91c-f27796ab40f4" (UID: "63972ba9-04aa-42e8-a91c-f27796ab40f4"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.422607 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-idp-0-file-data\") pod \"63972ba9-04aa-42e8-a91c-f27796ab40f4\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.422673 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-session\") pod \"63972ba9-04aa-42e8-a91c-f27796ab40f4\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.422729 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-ocp-branding-template\") pod \"63972ba9-04aa-42e8-a91c-f27796ab40f4\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.422777 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8npcc\" (UniqueName: \"kubernetes.io/projected/63972ba9-04aa-42e8-a91c-f27796ab40f4-kube-api-access-8npcc\") pod \"63972ba9-04aa-42e8-a91c-f27796ab40f4\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.422860 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-audit-policies\") pod \"63972ba9-04aa-42e8-a91c-f27796ab40f4\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.422914 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-template-provider-selection\") pod \"63972ba9-04aa-42e8-a91c-f27796ab40f4\" (UID: \"63972ba9-04aa-42e8-a91c-f27796ab40f4\") " Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.423516 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "63972ba9-04aa-42e8-a91c-f27796ab40f4" (UID: "63972ba9-04aa-42e8-a91c-f27796ab40f4"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.423544 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "63972ba9-04aa-42e8-a91c-f27796ab40f4" (UID: "63972ba9-04aa-42e8-a91c-f27796ab40f4"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.423726 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "63972ba9-04aa-42e8-a91c-f27796ab40f4" (UID: "63972ba9-04aa-42e8-a91c-f27796ab40f4"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.424111 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.424223 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.425284 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "63972ba9-04aa-42e8-a91c-f27796ab40f4" (UID: "63972ba9-04aa-42e8-a91c-f27796ab40f4"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.425465 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-router-certs\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.425541 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-service-ca\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.425626 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c6df39a7-5145-4f72-bd5e-2bef2077498c-audit-dir\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.425718 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-user-template-error\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.425809 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-user-template-login\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.425902 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.425959 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.426006 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.426058 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkkt6\" (UniqueName: \"kubernetes.io/projected/c6df39a7-5145-4f72-bd5e-2bef2077498c-kube-api-access-rkkt6\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.426173 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-session\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.426267 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.426337 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c6df39a7-5145-4f72-bd5e-2bef2077498c-audit-policies\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.426757 4802 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.426824 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.426853 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.426878 4802 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/63972ba9-04aa-42e8-a91c-f27796ab40f4-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.426901 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.426906 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.427994 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c6df39a7-5145-4f72-bd5e-2bef2077498c-audit-policies\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.430764 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "63972ba9-04aa-42e8-a91c-f27796ab40f4" (UID: "63972ba9-04aa-42e8-a91c-f27796ab40f4"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.430887 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c6df39a7-5145-4f72-bd5e-2bef2077498c-audit-dir\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.431233 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63972ba9-04aa-42e8-a91c-f27796ab40f4-kube-api-access-8npcc" (OuterVolumeSpecName: "kube-api-access-8npcc") pod "63972ba9-04aa-42e8-a91c-f27796ab40f4" (UID: "63972ba9-04aa-42e8-a91c-f27796ab40f4"). InnerVolumeSpecName "kube-api-access-8npcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.431398 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "63972ba9-04aa-42e8-a91c-f27796ab40f4" (UID: "63972ba9-04aa-42e8-a91c-f27796ab40f4"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.431499 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.431789 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "63972ba9-04aa-42e8-a91c-f27796ab40f4" (UID: "63972ba9-04aa-42e8-a91c-f27796ab40f4"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.431978 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-router-certs\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.432235 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "63972ba9-04aa-42e8-a91c-f27796ab40f4" (UID: "63972ba9-04aa-42e8-a91c-f27796ab40f4"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.432426 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-service-ca\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.432485 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.432940 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "63972ba9-04aa-42e8-a91c-f27796ab40f4" (UID: "63972ba9-04aa-42e8-a91c-f27796ab40f4"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.433732 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "63972ba9-04aa-42e8-a91c-f27796ab40f4" (UID: "63972ba9-04aa-42e8-a91c-f27796ab40f4"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.434100 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "63972ba9-04aa-42e8-a91c-f27796ab40f4" (UID: "63972ba9-04aa-42e8-a91c-f27796ab40f4"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.435068 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "63972ba9-04aa-42e8-a91c-f27796ab40f4" (UID: "63972ba9-04aa-42e8-a91c-f27796ab40f4"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.435988 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.436011 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-session\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.436514 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.437160 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-user-template-error\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.438474 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-user-template-login\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.438688 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c6df39a7-5145-4f72-bd5e-2bef2077498c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.452030 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkkt6\" (UniqueName: \"kubernetes.io/projected/c6df39a7-5145-4f72-bd5e-2bef2077498c-kube-api-access-rkkt6\") pod \"oauth-openshift-b6fcd9dcb-8kr95\" (UID: \"c6df39a7-5145-4f72-bd5e-2bef2077498c\") " pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.528920 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.528974 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.528990 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.529006 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.529022 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.529036 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.529051 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.529065 4802 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/63972ba9-04aa-42e8-a91c-f27796ab40f4-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.529080 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8npcc\" (UniqueName: \"kubernetes.io/projected/63972ba9-04aa-42e8-a91c-f27796ab40f4-kube-api-access-8npcc\") on node \"crc\" DevicePath \"\"" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.586103 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.596806 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sccr2"] Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.601003 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sccr2"] Nov 25 16:51:54 crc kubenswrapper[4802]: I1125 16:51:54.806562 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95"] Nov 25 16:51:55 crc kubenswrapper[4802]: I1125 16:51:55.264416 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" event={"ID":"c6df39a7-5145-4f72-bd5e-2bef2077498c","Type":"ContainerStarted","Data":"782b62b5acf1c1594b32f07bb26225a455c39df4a79e3f8feb14d12d46c232c0"} Nov 25 16:51:55 crc kubenswrapper[4802]: I1125 16:51:55.264743 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:51:55 crc kubenswrapper[4802]: I1125 16:51:55.264760 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" event={"ID":"c6df39a7-5145-4f72-bd5e-2bef2077498c","Type":"ContainerStarted","Data":"c3adab18ccf1dad83f693d0ae3cd4241432ebcac8c472fec782360ff6fd242f1"} Nov 25 16:51:55 crc kubenswrapper[4802]: I1125 16:51:55.292438 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" podStartSLOduration=27.292414462 podStartE2EDuration="27.292414462s" podCreationTimestamp="2025-11-25 16:51:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:51:55.286316955 +0000 UTC m=+298.430664141" watchObservedRunningTime="2025-11-25 16:51:55.292414462 +0000 UTC m=+298.436761648" Nov 25 16:51:55 crc kubenswrapper[4802]: I1125 16:51:55.511779 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63972ba9-04aa-42e8-a91c-f27796ab40f4" path="/var/lib/kubelet/pods/63972ba9-04aa-42e8-a91c-f27796ab40f4/volumes" Nov 25 16:51:55 crc kubenswrapper[4802]: I1125 16:51:55.925688 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-b6fcd9dcb-8kr95" Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.475627 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rtgv8"] Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.476834 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rtgv8" podUID="565aa173-3fac-4f4e-8fae-3fbe88184efb" containerName="registry-server" containerID="cri-o://ccf1b5358fa3e74a0152ef8ccd66173125f86b34474a6386b043cb45c1f1332a" gracePeriod=30 Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.483376 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r9nw2"] Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.483672 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r9nw2" podUID="9b19f412-72dd-4fe9-b0f1-3abc306653ae" containerName="registry-server" containerID="cri-o://687abc839c294e282339ef0bfeb71caeaa59d38925acff4513fec772540ec13f" gracePeriod=30 Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.500306 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cr9nk"] Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.500626 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" podUID="98e0e18f-d88b-416a-b18b-eaaea13c8e21" containerName="marketplace-operator" containerID="cri-o://919559eccc834acc5ef1603d3125b827234982db0f9ad6449fa9be5dfd9d26ae" gracePeriod=30 Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.520941 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-469c8"] Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.521704 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-469c8" podUID="8790f862-7ccc-4b14-8c99-1f64349079e0" containerName="registry-server" containerID="cri-o://b49b627216d874a034e39e50347a94efad8726d938111a0562a4918720c3093c" gracePeriod=30 Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.531103 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wsq47"] Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.531497 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wsq47" podUID="55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21" containerName="registry-server" containerID="cri-o://6feb3e15a0c77aa78855831a496db8635c09a073f5e22b439fc5a8d6bd0fafcb" gracePeriod=30 Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.538399 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rxhj4"] Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.539291 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rxhj4" Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.597232 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/90e4d7a1-f42a-4a46-9dc7-a8778f357bb7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rxhj4\" (UID: \"90e4d7a1-f42a-4a46-9dc7-a8778f357bb7\") " pod="openshift-marketplace/marketplace-operator-79b997595-rxhj4" Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.597332 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlx6m\" (UniqueName: \"kubernetes.io/projected/90e4d7a1-f42a-4a46-9dc7-a8778f357bb7-kube-api-access-qlx6m\") pod \"marketplace-operator-79b997595-rxhj4\" (UID: \"90e4d7a1-f42a-4a46-9dc7-a8778f357bb7\") " pod="openshift-marketplace/marketplace-operator-79b997595-rxhj4" Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.597358 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/90e4d7a1-f42a-4a46-9dc7-a8778f357bb7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rxhj4\" (UID: \"90e4d7a1-f42a-4a46-9dc7-a8778f357bb7\") " pod="openshift-marketplace/marketplace-operator-79b997595-rxhj4" Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.600940 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rxhj4"] Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.698452 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/90e4d7a1-f42a-4a46-9dc7-a8778f357bb7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rxhj4\" (UID: \"90e4d7a1-f42a-4a46-9dc7-a8778f357bb7\") " pod="openshift-marketplace/marketplace-operator-79b997595-rxhj4" Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.698523 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlx6m\" (UniqueName: \"kubernetes.io/projected/90e4d7a1-f42a-4a46-9dc7-a8778f357bb7-kube-api-access-qlx6m\") pod \"marketplace-operator-79b997595-rxhj4\" (UID: \"90e4d7a1-f42a-4a46-9dc7-a8778f357bb7\") " pod="openshift-marketplace/marketplace-operator-79b997595-rxhj4" Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.698557 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/90e4d7a1-f42a-4a46-9dc7-a8778f357bb7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rxhj4\" (UID: \"90e4d7a1-f42a-4a46-9dc7-a8778f357bb7\") " pod="openshift-marketplace/marketplace-operator-79b997595-rxhj4" Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.700653 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/90e4d7a1-f42a-4a46-9dc7-a8778f357bb7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rxhj4\" (UID: \"90e4d7a1-f42a-4a46-9dc7-a8778f357bb7\") " pod="openshift-marketplace/marketplace-operator-79b997595-rxhj4" Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.707276 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/90e4d7a1-f42a-4a46-9dc7-a8778f357bb7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rxhj4\" (UID: \"90e4d7a1-f42a-4a46-9dc7-a8778f357bb7\") " pod="openshift-marketplace/marketplace-operator-79b997595-rxhj4" Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.719359 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlx6m\" (UniqueName: \"kubernetes.io/projected/90e4d7a1-f42a-4a46-9dc7-a8778f357bb7-kube-api-access-qlx6m\") pod \"marketplace-operator-79b997595-rxhj4\" (UID: \"90e4d7a1-f42a-4a46-9dc7-a8778f357bb7\") " pod="openshift-marketplace/marketplace-operator-79b997595-rxhj4" Nov 25 16:52:13 crc kubenswrapper[4802]: I1125 16:52:13.877216 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rxhj4" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.095589 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rxhj4"] Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.400246 4802 generic.go:334] "Generic (PLEG): container finished" podID="9b19f412-72dd-4fe9-b0f1-3abc306653ae" containerID="687abc839c294e282339ef0bfeb71caeaa59d38925acff4513fec772540ec13f" exitCode=0 Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.400356 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9nw2" event={"ID":"9b19f412-72dd-4fe9-b0f1-3abc306653ae","Type":"ContainerDied","Data":"687abc839c294e282339ef0bfeb71caeaa59d38925acff4513fec772540ec13f"} Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.412319 4802 generic.go:334] "Generic (PLEG): container finished" podID="55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21" containerID="6feb3e15a0c77aa78855831a496db8635c09a073f5e22b439fc5a8d6bd0fafcb" exitCode=0 Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.412396 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsq47" event={"ID":"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21","Type":"ContainerDied","Data":"6feb3e15a0c77aa78855831a496db8635c09a073f5e22b439fc5a8d6bd0fafcb"} Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.414357 4802 generic.go:334] "Generic (PLEG): container finished" podID="8790f862-7ccc-4b14-8c99-1f64349079e0" containerID="b49b627216d874a034e39e50347a94efad8726d938111a0562a4918720c3093c" exitCode=0 Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.414414 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-469c8" event={"ID":"8790f862-7ccc-4b14-8c99-1f64349079e0","Type":"ContainerDied","Data":"b49b627216d874a034e39e50347a94efad8726d938111a0562a4918720c3093c"} Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.418743 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rxhj4" event={"ID":"90e4d7a1-f42a-4a46-9dc7-a8778f357bb7","Type":"ContainerStarted","Data":"f30b12a35931e52adbe480c5c5c740b90bf81a24e94c5b6176c875d0d7768aa4"} Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.418810 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rxhj4" event={"ID":"90e4d7a1-f42a-4a46-9dc7-a8778f357bb7","Type":"ContainerStarted","Data":"d6db145a1ca607447f690cb2797cb5e3be95749845c62a940099aa8ac0f70c5a"} Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.419531 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-rxhj4" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.423142 4802 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rxhj4 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.423187 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rxhj4" podUID="90e4d7a1-f42a-4a46-9dc7-a8778f357bb7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.423795 4802 generic.go:334] "Generic (PLEG): container finished" podID="98e0e18f-d88b-416a-b18b-eaaea13c8e21" containerID="919559eccc834acc5ef1603d3125b827234982db0f9ad6449fa9be5dfd9d26ae" exitCode=0 Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.423856 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" event={"ID":"98e0e18f-d88b-416a-b18b-eaaea13c8e21","Type":"ContainerDied","Data":"919559eccc834acc5ef1603d3125b827234982db0f9ad6449fa9be5dfd9d26ae"} Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.428669 4802 generic.go:334] "Generic (PLEG): container finished" podID="565aa173-3fac-4f4e-8fae-3fbe88184efb" containerID="ccf1b5358fa3e74a0152ef8ccd66173125f86b34474a6386b043cb45c1f1332a" exitCode=0 Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.428701 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rtgv8" event={"ID":"565aa173-3fac-4f4e-8fae-3fbe88184efb","Type":"ContainerDied","Data":"ccf1b5358fa3e74a0152ef8ccd66173125f86b34474a6386b043cb45c1f1332a"} Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.437026 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-rxhj4" podStartSLOduration=1.437004441 podStartE2EDuration="1.437004441s" podCreationTimestamp="2025-11-25 16:52:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:52:14.434743034 +0000 UTC m=+317.579090240" watchObservedRunningTime="2025-11-25 16:52:14.437004441 +0000 UTC m=+317.581351627" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.559646 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.612630 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w29tv\" (UniqueName: \"kubernetes.io/projected/98e0e18f-d88b-416a-b18b-eaaea13c8e21-kube-api-access-w29tv\") pod \"98e0e18f-d88b-416a-b18b-eaaea13c8e21\" (UID: \"98e0e18f-d88b-416a-b18b-eaaea13c8e21\") " Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.612689 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98e0e18f-d88b-416a-b18b-eaaea13c8e21-marketplace-trusted-ca\") pod \"98e0e18f-d88b-416a-b18b-eaaea13c8e21\" (UID: \"98e0e18f-d88b-416a-b18b-eaaea13c8e21\") " Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.612788 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/98e0e18f-d88b-416a-b18b-eaaea13c8e21-marketplace-operator-metrics\") pod \"98e0e18f-d88b-416a-b18b-eaaea13c8e21\" (UID: \"98e0e18f-d88b-416a-b18b-eaaea13c8e21\") " Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.616596 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98e0e18f-d88b-416a-b18b-eaaea13c8e21-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "98e0e18f-d88b-416a-b18b-eaaea13c8e21" (UID: "98e0e18f-d88b-416a-b18b-eaaea13c8e21"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.626438 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98e0e18f-d88b-416a-b18b-eaaea13c8e21-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "98e0e18f-d88b-416a-b18b-eaaea13c8e21" (UID: "98e0e18f-d88b-416a-b18b-eaaea13c8e21"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.637363 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98e0e18f-d88b-416a-b18b-eaaea13c8e21-kube-api-access-w29tv" (OuterVolumeSpecName: "kube-api-access-w29tv") pod "98e0e18f-d88b-416a-b18b-eaaea13c8e21" (UID: "98e0e18f-d88b-416a-b18b-eaaea13c8e21"). InnerVolumeSpecName "kube-api-access-w29tv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.714375 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w29tv\" (UniqueName: \"kubernetes.io/projected/98e0e18f-d88b-416a-b18b-eaaea13c8e21-kube-api-access-w29tv\") on node \"crc\" DevicePath \"\"" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.714411 4802 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98e0e18f-d88b-416a-b18b-eaaea13c8e21-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.714420 4802 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/98e0e18f-d88b-416a-b18b-eaaea13c8e21-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.722478 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rtgv8" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.726709 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r9nw2" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.732631 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wsq47" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.743420 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-469c8" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.814982 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b19f412-72dd-4fe9-b0f1-3abc306653ae-utilities\") pod \"9b19f412-72dd-4fe9-b0f1-3abc306653ae\" (UID: \"9b19f412-72dd-4fe9-b0f1-3abc306653ae\") " Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.815040 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/565aa173-3fac-4f4e-8fae-3fbe88184efb-catalog-content\") pod \"565aa173-3fac-4f4e-8fae-3fbe88184efb\" (UID: \"565aa173-3fac-4f4e-8fae-3fbe88184efb\") " Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.815070 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/565aa173-3fac-4f4e-8fae-3fbe88184efb-utilities\") pod \"565aa173-3fac-4f4e-8fae-3fbe88184efb\" (UID: \"565aa173-3fac-4f4e-8fae-3fbe88184efb\") " Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.815100 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ckk8\" (UniqueName: \"kubernetes.io/projected/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21-kube-api-access-4ckk8\") pod \"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21\" (UID: \"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21\") " Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.815209 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8790f862-7ccc-4b14-8c99-1f64349079e0-utilities\") pod \"8790f862-7ccc-4b14-8c99-1f64349079e0\" (UID: \"8790f862-7ccc-4b14-8c99-1f64349079e0\") " Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.815242 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b19f412-72dd-4fe9-b0f1-3abc306653ae-catalog-content\") pod \"9b19f412-72dd-4fe9-b0f1-3abc306653ae\" (UID: \"9b19f412-72dd-4fe9-b0f1-3abc306653ae\") " Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.815292 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21-catalog-content\") pod \"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21\" (UID: \"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21\") " Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.815318 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr8ld\" (UniqueName: \"kubernetes.io/projected/8790f862-7ccc-4b14-8c99-1f64349079e0-kube-api-access-sr8ld\") pod \"8790f862-7ccc-4b14-8c99-1f64349079e0\" (UID: \"8790f862-7ccc-4b14-8c99-1f64349079e0\") " Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.815338 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21-utilities\") pod \"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21\" (UID: \"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21\") " Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.815368 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qm7mf\" (UniqueName: \"kubernetes.io/projected/9b19f412-72dd-4fe9-b0f1-3abc306653ae-kube-api-access-qm7mf\") pod \"9b19f412-72dd-4fe9-b0f1-3abc306653ae\" (UID: \"9b19f412-72dd-4fe9-b0f1-3abc306653ae\") " Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.815424 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4h6dr\" (UniqueName: \"kubernetes.io/projected/565aa173-3fac-4f4e-8fae-3fbe88184efb-kube-api-access-4h6dr\") pod \"565aa173-3fac-4f4e-8fae-3fbe88184efb\" (UID: \"565aa173-3fac-4f4e-8fae-3fbe88184efb\") " Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.815457 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8790f862-7ccc-4b14-8c99-1f64349079e0-catalog-content\") pod \"8790f862-7ccc-4b14-8c99-1f64349079e0\" (UID: \"8790f862-7ccc-4b14-8c99-1f64349079e0\") " Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.815800 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b19f412-72dd-4fe9-b0f1-3abc306653ae-utilities" (OuterVolumeSpecName: "utilities") pod "9b19f412-72dd-4fe9-b0f1-3abc306653ae" (UID: "9b19f412-72dd-4fe9-b0f1-3abc306653ae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.816723 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21-utilities" (OuterVolumeSpecName: "utilities") pod "55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21" (UID: "55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.819217 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8790f862-7ccc-4b14-8c99-1f64349079e0-utilities" (OuterVolumeSpecName: "utilities") pod "8790f862-7ccc-4b14-8c99-1f64349079e0" (UID: "8790f862-7ccc-4b14-8c99-1f64349079e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.820069 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8790f862-7ccc-4b14-8c99-1f64349079e0-kube-api-access-sr8ld" (OuterVolumeSpecName: "kube-api-access-sr8ld") pod "8790f862-7ccc-4b14-8c99-1f64349079e0" (UID: "8790f862-7ccc-4b14-8c99-1f64349079e0"). InnerVolumeSpecName "kube-api-access-sr8ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.820581 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/565aa173-3fac-4f4e-8fae-3fbe88184efb-utilities" (OuterVolumeSpecName: "utilities") pod "565aa173-3fac-4f4e-8fae-3fbe88184efb" (UID: "565aa173-3fac-4f4e-8fae-3fbe88184efb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.820612 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b19f412-72dd-4fe9-b0f1-3abc306653ae-kube-api-access-qm7mf" (OuterVolumeSpecName: "kube-api-access-qm7mf") pod "9b19f412-72dd-4fe9-b0f1-3abc306653ae" (UID: "9b19f412-72dd-4fe9-b0f1-3abc306653ae"). InnerVolumeSpecName "kube-api-access-qm7mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.822193 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21-kube-api-access-4ckk8" (OuterVolumeSpecName: "kube-api-access-4ckk8") pod "55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21" (UID: "55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21"). InnerVolumeSpecName "kube-api-access-4ckk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.824761 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/565aa173-3fac-4f4e-8fae-3fbe88184efb-kube-api-access-4h6dr" (OuterVolumeSpecName: "kube-api-access-4h6dr") pod "565aa173-3fac-4f4e-8fae-3fbe88184efb" (UID: "565aa173-3fac-4f4e-8fae-3fbe88184efb"). InnerVolumeSpecName "kube-api-access-4h6dr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.857817 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8790f862-7ccc-4b14-8c99-1f64349079e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8790f862-7ccc-4b14-8c99-1f64349079e0" (UID: "8790f862-7ccc-4b14-8c99-1f64349079e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.883031 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/565aa173-3fac-4f4e-8fae-3fbe88184efb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "565aa173-3fac-4f4e-8fae-3fbe88184efb" (UID: "565aa173-3fac-4f4e-8fae-3fbe88184efb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.888883 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b19f412-72dd-4fe9-b0f1-3abc306653ae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9b19f412-72dd-4fe9-b0f1-3abc306653ae" (UID: "9b19f412-72dd-4fe9-b0f1-3abc306653ae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.916418 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qm7mf\" (UniqueName: \"kubernetes.io/projected/9b19f412-72dd-4fe9-b0f1-3abc306653ae-kube-api-access-qm7mf\") on node \"crc\" DevicePath \"\"" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.916455 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4h6dr\" (UniqueName: \"kubernetes.io/projected/565aa173-3fac-4f4e-8fae-3fbe88184efb-kube-api-access-4h6dr\") on node \"crc\" DevicePath \"\"" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.916466 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8790f862-7ccc-4b14-8c99-1f64349079e0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.916475 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b19f412-72dd-4fe9-b0f1-3abc306653ae-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.916484 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/565aa173-3fac-4f4e-8fae-3fbe88184efb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.916492 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/565aa173-3fac-4f4e-8fae-3fbe88184efb-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.916500 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ckk8\" (UniqueName: \"kubernetes.io/projected/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21-kube-api-access-4ckk8\") on node \"crc\" DevicePath \"\"" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.916507 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8790f862-7ccc-4b14-8c99-1f64349079e0-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.916517 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b19f412-72dd-4fe9-b0f1-3abc306653ae-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.916525 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr8ld\" (UniqueName: \"kubernetes.io/projected/8790f862-7ccc-4b14-8c99-1f64349079e0-kube-api-access-sr8ld\") on node \"crc\" DevicePath \"\"" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.916534 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:52:14 crc kubenswrapper[4802]: I1125 16:52:14.930279 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21" (UID: "55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.018016 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.444526 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" event={"ID":"98e0e18f-d88b-416a-b18b-eaaea13c8e21","Type":"ContainerDied","Data":"e7c125bfb5eaf1161b88597f542893fcac84d0ce40086a4b12ad7c1977ed48cb"} Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.444600 4802 scope.go:117] "RemoveContainer" containerID="919559eccc834acc5ef1603d3125b827234982db0f9ad6449fa9be5dfd9d26ae" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.444538 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cr9nk" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.450802 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rtgv8" event={"ID":"565aa173-3fac-4f4e-8fae-3fbe88184efb","Type":"ContainerDied","Data":"404aeaf695d8748ce2de721578e021d59ff690576206fb5571ec41c5807bcbdb"} Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.451077 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rtgv8" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.455361 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9nw2" event={"ID":"9b19f412-72dd-4fe9-b0f1-3abc306653ae","Type":"ContainerDied","Data":"04f5263325584e02baed24b8a86b968ad3334c730b423247b61d707bffc44983"} Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.455601 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r9nw2" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.466377 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wsq47" event={"ID":"55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21","Type":"ContainerDied","Data":"cd0e0ac1919ff451dfe62d09dd9b8169eecda46a1312f0a13824aa8ca96a8aba"} Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.467009 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wsq47" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.468042 4802 scope.go:117] "RemoveContainer" containerID="ccf1b5358fa3e74a0152ef8ccd66173125f86b34474a6386b043cb45c1f1332a" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.469874 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-469c8" event={"ID":"8790f862-7ccc-4b14-8c99-1f64349079e0","Type":"ContainerDied","Data":"3102be8c615409385fbccc78cdee471dd76eb9e6fbfa1210f19d86d6e87feaef"} Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.469915 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-469c8" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.473505 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-rxhj4" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.490330 4802 scope.go:117] "RemoveContainer" containerID="7e4823bc4a29a760bd06c65226babca4c32332c8bd477fcf7f85d3406c3e78ee" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.543096 4802 scope.go:117] "RemoveContainer" containerID="7c845f5220c08bf1df0f8ed5cec4f02229391b92ad53c79d8566ce5b1b3c262e" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.547554 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rtgv8"] Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.547588 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rtgv8"] Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.547622 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wsq47"] Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.553223 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wsq47"] Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.557226 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r9nw2"] Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.559789 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r9nw2"] Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.564841 4802 scope.go:117] "RemoveContainer" containerID="687abc839c294e282339ef0bfeb71caeaa59d38925acff4513fec772540ec13f" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.566256 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cr9nk"] Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.569312 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cr9nk"] Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.572236 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-469c8"] Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.574544 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-469c8"] Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.578587 4802 scope.go:117] "RemoveContainer" containerID="e1767772901a233e33ef62e16d14a3fc767a43eac511a4d01dd4f8c98542cc20" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.594311 4802 scope.go:117] "RemoveContainer" containerID="f0614a9f2967c4648c9aa958d07ef7d86aef14fa6dbbee7ee8ddd5808040ec68" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.626842 4802 scope.go:117] "RemoveContainer" containerID="6feb3e15a0c77aa78855831a496db8635c09a073f5e22b439fc5a8d6bd0fafcb" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.641232 4802 scope.go:117] "RemoveContainer" containerID="2ff38bf38de72c6caeaa5608f69f032884570e0171406fa14892004c20d88b53" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.663274 4802 scope.go:117] "RemoveContainer" containerID="ed1e7d350328f21c0c42f7a18af9596b5e7fcc77c2e114918e9152d87ff88c3a" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.677040 4802 scope.go:117] "RemoveContainer" containerID="b49b627216d874a034e39e50347a94efad8726d938111a0562a4918720c3093c" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.692427 4802 scope.go:117] "RemoveContainer" containerID="98949a755bbc45a3e7660dfc427e538d79817ac5678486ce554d0814e3f66253" Nov 25 16:52:15 crc kubenswrapper[4802]: I1125 16:52:15.707959 4802 scope.go:117] "RemoveContainer" containerID="d413871f5ffe933b3d095fd476deac62abe1f70847ee024955eae178d759f358" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.518040 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21" path="/var/lib/kubelet/pods/55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21/volumes" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.519819 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="565aa173-3fac-4f4e-8fae-3fbe88184efb" path="/var/lib/kubelet/pods/565aa173-3fac-4f4e-8fae-3fbe88184efb/volumes" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.521079 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8790f862-7ccc-4b14-8c99-1f64349079e0" path="/var/lib/kubelet/pods/8790f862-7ccc-4b14-8c99-1f64349079e0/volumes" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.523252 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98e0e18f-d88b-416a-b18b-eaaea13c8e21" path="/var/lib/kubelet/pods/98e0e18f-d88b-416a-b18b-eaaea13c8e21/volumes" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.524176 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b19f412-72dd-4fe9-b0f1-3abc306653ae" path="/var/lib/kubelet/pods/9b19f412-72dd-4fe9-b0f1-3abc306653ae/volumes" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696109 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rslrk"] Nov 25 16:52:17 crc kubenswrapper[4802]: E1125 16:52:17.696412 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21" containerName="extract-utilities" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696429 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21" containerName="extract-utilities" Nov 25 16:52:17 crc kubenswrapper[4802]: E1125 16:52:17.696445 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="565aa173-3fac-4f4e-8fae-3fbe88184efb" containerName="extract-content" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696455 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="565aa173-3fac-4f4e-8fae-3fbe88184efb" containerName="extract-content" Nov 25 16:52:17 crc kubenswrapper[4802]: E1125 16:52:17.696466 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21" containerName="extract-content" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696477 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21" containerName="extract-content" Nov 25 16:52:17 crc kubenswrapper[4802]: E1125 16:52:17.696499 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b19f412-72dd-4fe9-b0f1-3abc306653ae" containerName="registry-server" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696508 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b19f412-72dd-4fe9-b0f1-3abc306653ae" containerName="registry-server" Nov 25 16:52:17 crc kubenswrapper[4802]: E1125 16:52:17.696520 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="565aa173-3fac-4f4e-8fae-3fbe88184efb" containerName="registry-server" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696527 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="565aa173-3fac-4f4e-8fae-3fbe88184efb" containerName="registry-server" Nov 25 16:52:17 crc kubenswrapper[4802]: E1125 16:52:17.696539 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b19f412-72dd-4fe9-b0f1-3abc306653ae" containerName="extract-utilities" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696547 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b19f412-72dd-4fe9-b0f1-3abc306653ae" containerName="extract-utilities" Nov 25 16:52:17 crc kubenswrapper[4802]: E1125 16:52:17.696557 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98e0e18f-d88b-416a-b18b-eaaea13c8e21" containerName="marketplace-operator" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696565 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="98e0e18f-d88b-416a-b18b-eaaea13c8e21" containerName="marketplace-operator" Nov 25 16:52:17 crc kubenswrapper[4802]: E1125 16:52:17.696575 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8790f862-7ccc-4b14-8c99-1f64349079e0" containerName="extract-content" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696583 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="8790f862-7ccc-4b14-8c99-1f64349079e0" containerName="extract-content" Nov 25 16:52:17 crc kubenswrapper[4802]: E1125 16:52:17.696595 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b19f412-72dd-4fe9-b0f1-3abc306653ae" containerName="extract-content" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696603 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b19f412-72dd-4fe9-b0f1-3abc306653ae" containerName="extract-content" Nov 25 16:52:17 crc kubenswrapper[4802]: E1125 16:52:17.696613 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="565aa173-3fac-4f4e-8fae-3fbe88184efb" containerName="extract-utilities" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696621 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="565aa173-3fac-4f4e-8fae-3fbe88184efb" containerName="extract-utilities" Nov 25 16:52:17 crc kubenswrapper[4802]: E1125 16:52:17.696631 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8790f862-7ccc-4b14-8c99-1f64349079e0" containerName="extract-utilities" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696640 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="8790f862-7ccc-4b14-8c99-1f64349079e0" containerName="extract-utilities" Nov 25 16:52:17 crc kubenswrapper[4802]: E1125 16:52:17.696649 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21" containerName="registry-server" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696657 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21" containerName="registry-server" Nov 25 16:52:17 crc kubenswrapper[4802]: E1125 16:52:17.696671 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8790f862-7ccc-4b14-8c99-1f64349079e0" containerName="registry-server" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696679 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="8790f862-7ccc-4b14-8c99-1f64349079e0" containerName="registry-server" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696786 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="98e0e18f-d88b-416a-b18b-eaaea13c8e21" containerName="marketplace-operator" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696802 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="565aa173-3fac-4f4e-8fae-3fbe88184efb" containerName="registry-server" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696818 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b19f412-72dd-4fe9-b0f1-3abc306653ae" containerName="registry-server" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696828 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="8790f862-7ccc-4b14-8c99-1f64349079e0" containerName="registry-server" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.696838 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="55ae2bee-8b1d-46a6-a6c2-76c97f8bdf21" containerName="registry-server" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.697719 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rslrk" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.699815 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.711464 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rslrk"] Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.751205 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd-catalog-content\") pod \"certified-operators-rslrk\" (UID: \"6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd\") " pod="openshift-marketplace/certified-operators-rslrk" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.751559 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd-utilities\") pod \"certified-operators-rslrk\" (UID: \"6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd\") " pod="openshift-marketplace/certified-operators-rslrk" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.752256 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgqz7\" (UniqueName: \"kubernetes.io/projected/6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd-kube-api-access-vgqz7\") pod \"certified-operators-rslrk\" (UID: \"6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd\") " pod="openshift-marketplace/certified-operators-rslrk" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.853015 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgqz7\" (UniqueName: \"kubernetes.io/projected/6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd-kube-api-access-vgqz7\") pod \"certified-operators-rslrk\" (UID: \"6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd\") " pod="openshift-marketplace/certified-operators-rslrk" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.853086 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd-catalog-content\") pod \"certified-operators-rslrk\" (UID: \"6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd\") " pod="openshift-marketplace/certified-operators-rslrk" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.853108 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd-utilities\") pod \"certified-operators-rslrk\" (UID: \"6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd\") " pod="openshift-marketplace/certified-operators-rslrk" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.853632 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd-catalog-content\") pod \"certified-operators-rslrk\" (UID: \"6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd\") " pod="openshift-marketplace/certified-operators-rslrk" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.853742 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd-utilities\") pod \"certified-operators-rslrk\" (UID: \"6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd\") " pod="openshift-marketplace/certified-operators-rslrk" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.872464 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgqz7\" (UniqueName: \"kubernetes.io/projected/6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd-kube-api-access-vgqz7\") pod \"certified-operators-rslrk\" (UID: \"6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd\") " pod="openshift-marketplace/certified-operators-rslrk" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.894331 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mlz5g"] Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.895771 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mlz5g" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.898276 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.903179 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mlz5g"] Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.954355 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wknsx\" (UniqueName: \"kubernetes.io/projected/ba359221-8966-436b-b846-4c5ac2f1767e-kube-api-access-wknsx\") pod \"community-operators-mlz5g\" (UID: \"ba359221-8966-436b-b846-4c5ac2f1767e\") " pod="openshift-marketplace/community-operators-mlz5g" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.954416 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba359221-8966-436b-b846-4c5ac2f1767e-catalog-content\") pod \"community-operators-mlz5g\" (UID: \"ba359221-8966-436b-b846-4c5ac2f1767e\") " pod="openshift-marketplace/community-operators-mlz5g" Nov 25 16:52:17 crc kubenswrapper[4802]: I1125 16:52:17.954437 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba359221-8966-436b-b846-4c5ac2f1767e-utilities\") pod \"community-operators-mlz5g\" (UID: \"ba359221-8966-436b-b846-4c5ac2f1767e\") " pod="openshift-marketplace/community-operators-mlz5g" Nov 25 16:52:18 crc kubenswrapper[4802]: I1125 16:52:18.033081 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rslrk" Nov 25 16:52:18 crc kubenswrapper[4802]: I1125 16:52:18.055596 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba359221-8966-436b-b846-4c5ac2f1767e-catalog-content\") pod \"community-operators-mlz5g\" (UID: \"ba359221-8966-436b-b846-4c5ac2f1767e\") " pod="openshift-marketplace/community-operators-mlz5g" Nov 25 16:52:18 crc kubenswrapper[4802]: I1125 16:52:18.056060 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba359221-8966-436b-b846-4c5ac2f1767e-utilities\") pod \"community-operators-mlz5g\" (UID: \"ba359221-8966-436b-b846-4c5ac2f1767e\") " pod="openshift-marketplace/community-operators-mlz5g" Nov 25 16:52:18 crc kubenswrapper[4802]: I1125 16:52:18.056079 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba359221-8966-436b-b846-4c5ac2f1767e-catalog-content\") pod \"community-operators-mlz5g\" (UID: \"ba359221-8966-436b-b846-4c5ac2f1767e\") " pod="openshift-marketplace/community-operators-mlz5g" Nov 25 16:52:18 crc kubenswrapper[4802]: I1125 16:52:18.056141 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wknsx\" (UniqueName: \"kubernetes.io/projected/ba359221-8966-436b-b846-4c5ac2f1767e-kube-api-access-wknsx\") pod \"community-operators-mlz5g\" (UID: \"ba359221-8966-436b-b846-4c5ac2f1767e\") " pod="openshift-marketplace/community-operators-mlz5g" Nov 25 16:52:18 crc kubenswrapper[4802]: I1125 16:52:18.056341 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba359221-8966-436b-b846-4c5ac2f1767e-utilities\") pod \"community-operators-mlz5g\" (UID: \"ba359221-8966-436b-b846-4c5ac2f1767e\") " pod="openshift-marketplace/community-operators-mlz5g" Nov 25 16:52:18 crc kubenswrapper[4802]: I1125 16:52:18.085949 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wknsx\" (UniqueName: \"kubernetes.io/projected/ba359221-8966-436b-b846-4c5ac2f1767e-kube-api-access-wknsx\") pod \"community-operators-mlz5g\" (UID: \"ba359221-8966-436b-b846-4c5ac2f1767e\") " pod="openshift-marketplace/community-operators-mlz5g" Nov 25 16:52:18 crc kubenswrapper[4802]: I1125 16:52:18.218156 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mlz5g" Nov 25 16:52:18 crc kubenswrapper[4802]: I1125 16:52:18.225525 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rslrk"] Nov 25 16:52:18 crc kubenswrapper[4802]: W1125 16:52:18.242089 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6afc1234_7d7b_4da1_a2c3_62ee8e50b2bd.slice/crio-fffba5d1657223eaf65e767d5607e4643a0fe8bd93f1e4d7a713ad856880204a WatchSource:0}: Error finding container fffba5d1657223eaf65e767d5607e4643a0fe8bd93f1e4d7a713ad856880204a: Status 404 returned error can't find the container with id fffba5d1657223eaf65e767d5607e4643a0fe8bd93f1e4d7a713ad856880204a Nov 25 16:52:18 crc kubenswrapper[4802]: I1125 16:52:18.422554 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mlz5g"] Nov 25 16:52:18 crc kubenswrapper[4802]: W1125 16:52:18.474035 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba359221_8966_436b_b846_4c5ac2f1767e.slice/crio-b0595f1080f1003da08553d1bd05b628ba45f9360e257025845f1305ac78977b WatchSource:0}: Error finding container b0595f1080f1003da08553d1bd05b628ba45f9360e257025845f1305ac78977b: Status 404 returned error can't find the container with id b0595f1080f1003da08553d1bd05b628ba45f9360e257025845f1305ac78977b Nov 25 16:52:18 crc kubenswrapper[4802]: I1125 16:52:18.495620 4802 generic.go:334] "Generic (PLEG): container finished" podID="6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd" containerID="6746659c70486b3b97232ee71eee7c529bc4760e9c480c2c5dd79ff81af5de85" exitCode=0 Nov 25 16:52:18 crc kubenswrapper[4802]: I1125 16:52:18.495683 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rslrk" event={"ID":"6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd","Type":"ContainerDied","Data":"6746659c70486b3b97232ee71eee7c529bc4760e9c480c2c5dd79ff81af5de85"} Nov 25 16:52:18 crc kubenswrapper[4802]: I1125 16:52:18.495746 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rslrk" event={"ID":"6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd","Type":"ContainerStarted","Data":"fffba5d1657223eaf65e767d5607e4643a0fe8bd93f1e4d7a713ad856880204a"} Nov 25 16:52:18 crc kubenswrapper[4802]: I1125 16:52:18.496834 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlz5g" event={"ID":"ba359221-8966-436b-b846-4c5ac2f1767e","Type":"ContainerStarted","Data":"b0595f1080f1003da08553d1bd05b628ba45f9360e257025845f1305ac78977b"} Nov 25 16:52:19 crc kubenswrapper[4802]: I1125 16:52:19.508505 4802 generic.go:334] "Generic (PLEG): container finished" podID="ba359221-8966-436b-b846-4c5ac2f1767e" containerID="6eb6a4338397f6ded018f9591f512a6d593dfe6f084c88962fc56d407fcc9a7a" exitCode=0 Nov 25 16:52:19 crc kubenswrapper[4802]: I1125 16:52:19.519715 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlz5g" event={"ID":"ba359221-8966-436b-b846-4c5ac2f1767e","Type":"ContainerDied","Data":"6eb6a4338397f6ded018f9591f512a6d593dfe6f084c88962fc56d407fcc9a7a"} Nov 25 16:52:19 crc kubenswrapper[4802]: I1125 16:52:19.519760 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rslrk" event={"ID":"6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd","Type":"ContainerStarted","Data":"2a9d4a13575cf894588954e802f53e49dfb396cc6ad6e408c6234777dfe3837a"} Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.105534 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f5llz"] Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.106727 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5llz" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.110072 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.110315 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5llz"] Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.179982 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5-utilities\") pod \"redhat-marketplace-f5llz\" (UID: \"b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5\") " pod="openshift-marketplace/redhat-marketplace-f5llz" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.180099 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5-catalog-content\") pod \"redhat-marketplace-f5llz\" (UID: \"b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5\") " pod="openshift-marketplace/redhat-marketplace-f5llz" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.180204 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nl4df\" (UniqueName: \"kubernetes.io/projected/b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5-kube-api-access-nl4df\") pod \"redhat-marketplace-f5llz\" (UID: \"b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5\") " pod="openshift-marketplace/redhat-marketplace-f5llz" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.281311 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5-utilities\") pod \"redhat-marketplace-f5llz\" (UID: \"b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5\") " pod="openshift-marketplace/redhat-marketplace-f5llz" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.281392 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5-catalog-content\") pod \"redhat-marketplace-f5llz\" (UID: \"b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5\") " pod="openshift-marketplace/redhat-marketplace-f5llz" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.281427 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nl4df\" (UniqueName: \"kubernetes.io/projected/b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5-kube-api-access-nl4df\") pod \"redhat-marketplace-f5llz\" (UID: \"b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5\") " pod="openshift-marketplace/redhat-marketplace-f5llz" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.282079 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5-catalog-content\") pod \"redhat-marketplace-f5llz\" (UID: \"b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5\") " pod="openshift-marketplace/redhat-marketplace-f5llz" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.282163 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5-utilities\") pod \"redhat-marketplace-f5llz\" (UID: \"b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5\") " pod="openshift-marketplace/redhat-marketplace-f5llz" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.301341 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kw428"] Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.302751 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kw428" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.307602 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.314019 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kw428"] Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.326905 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nl4df\" (UniqueName: \"kubernetes.io/projected/b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5-kube-api-access-nl4df\") pod \"redhat-marketplace-f5llz\" (UID: \"b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5\") " pod="openshift-marketplace/redhat-marketplace-f5llz" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.383306 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/495e0a1d-2dd0-4069-a9dd-6637b835a35a-utilities\") pod \"redhat-operators-kw428\" (UID: \"495e0a1d-2dd0-4069-a9dd-6637b835a35a\") " pod="openshift-marketplace/redhat-operators-kw428" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.383414 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkj5k\" (UniqueName: \"kubernetes.io/projected/495e0a1d-2dd0-4069-a9dd-6637b835a35a-kube-api-access-kkj5k\") pod \"redhat-operators-kw428\" (UID: \"495e0a1d-2dd0-4069-a9dd-6637b835a35a\") " pod="openshift-marketplace/redhat-operators-kw428" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.383459 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/495e0a1d-2dd0-4069-a9dd-6637b835a35a-catalog-content\") pod \"redhat-operators-kw428\" (UID: \"495e0a1d-2dd0-4069-a9dd-6637b835a35a\") " pod="openshift-marketplace/redhat-operators-kw428" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.476550 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5llz" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.485086 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/495e0a1d-2dd0-4069-a9dd-6637b835a35a-utilities\") pod \"redhat-operators-kw428\" (UID: \"495e0a1d-2dd0-4069-a9dd-6637b835a35a\") " pod="openshift-marketplace/redhat-operators-kw428" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.485234 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkj5k\" (UniqueName: \"kubernetes.io/projected/495e0a1d-2dd0-4069-a9dd-6637b835a35a-kube-api-access-kkj5k\") pod \"redhat-operators-kw428\" (UID: \"495e0a1d-2dd0-4069-a9dd-6637b835a35a\") " pod="openshift-marketplace/redhat-operators-kw428" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.485281 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/495e0a1d-2dd0-4069-a9dd-6637b835a35a-catalog-content\") pod \"redhat-operators-kw428\" (UID: \"495e0a1d-2dd0-4069-a9dd-6637b835a35a\") " pod="openshift-marketplace/redhat-operators-kw428" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.486011 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/495e0a1d-2dd0-4069-a9dd-6637b835a35a-catalog-content\") pod \"redhat-operators-kw428\" (UID: \"495e0a1d-2dd0-4069-a9dd-6637b835a35a\") " pod="openshift-marketplace/redhat-operators-kw428" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.486254 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/495e0a1d-2dd0-4069-a9dd-6637b835a35a-utilities\") pod \"redhat-operators-kw428\" (UID: \"495e0a1d-2dd0-4069-a9dd-6637b835a35a\") " pod="openshift-marketplace/redhat-operators-kw428" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.503500 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkj5k\" (UniqueName: \"kubernetes.io/projected/495e0a1d-2dd0-4069-a9dd-6637b835a35a-kube-api-access-kkj5k\") pod \"redhat-operators-kw428\" (UID: \"495e0a1d-2dd0-4069-a9dd-6637b835a35a\") " pod="openshift-marketplace/redhat-operators-kw428" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.526012 4802 generic.go:334] "Generic (PLEG): container finished" podID="6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd" containerID="2a9d4a13575cf894588954e802f53e49dfb396cc6ad6e408c6234777dfe3837a" exitCode=0 Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.527567 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rslrk" event={"ID":"6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd","Type":"ContainerDied","Data":"2a9d4a13575cf894588954e802f53e49dfb396cc6ad6e408c6234777dfe3837a"} Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.531841 4802 generic.go:334] "Generic (PLEG): container finished" podID="ba359221-8966-436b-b846-4c5ac2f1767e" containerID="91507a62427e1adcf2327f735569343c56e1425027361eae3f11495d01bb086e" exitCode=0 Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.531952 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlz5g" event={"ID":"ba359221-8966-436b-b846-4c5ac2f1767e","Type":"ContainerDied","Data":"91507a62427e1adcf2327f735569343c56e1425027361eae3f11495d01bb086e"} Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.652574 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kw428" Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.690630 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5llz"] Nov 25 16:52:20 crc kubenswrapper[4802]: W1125 16:52:20.694979 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2ae61bd_15c2_47c4_9cf0_8161fae2f3d5.slice/crio-2ffd8c0afeff16333d67fe9efe2c7ef21f86d94fe8118363190d8a608f96621a WatchSource:0}: Error finding container 2ffd8c0afeff16333d67fe9efe2c7ef21f86d94fe8118363190d8a608f96621a: Status 404 returned error can't find the container with id 2ffd8c0afeff16333d67fe9efe2c7ef21f86d94fe8118363190d8a608f96621a Nov 25 16:52:20 crc kubenswrapper[4802]: I1125 16:52:20.860761 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kw428"] Nov 25 16:52:20 crc kubenswrapper[4802]: W1125 16:52:20.937762 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod495e0a1d_2dd0_4069_a9dd_6637b835a35a.slice/crio-a0a16d9a239bfb7415b85a13b85b5b791b2fbdd4b43576c693e9790ffeec17b1 WatchSource:0}: Error finding container a0a16d9a239bfb7415b85a13b85b5b791b2fbdd4b43576c693e9790ffeec17b1: Status 404 returned error can't find the container with id a0a16d9a239bfb7415b85a13b85b5b791b2fbdd4b43576c693e9790ffeec17b1 Nov 25 16:52:21 crc kubenswrapper[4802]: I1125 16:52:21.541748 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rslrk" event={"ID":"6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd","Type":"ContainerStarted","Data":"d8ad006d8e78bf1e4ff3279e2c2d9ca480901053e818497251925de52f3322ec"} Nov 25 16:52:21 crc kubenswrapper[4802]: I1125 16:52:21.545796 4802 generic.go:334] "Generic (PLEG): container finished" podID="b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5" containerID="f02d161b7cc818970e83ea0bca3f9b004c81279122bff79d269eb59e99e5d886" exitCode=0 Nov 25 16:52:21 crc kubenswrapper[4802]: I1125 16:52:21.545852 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5llz" event={"ID":"b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5","Type":"ContainerDied","Data":"f02d161b7cc818970e83ea0bca3f9b004c81279122bff79d269eb59e99e5d886"} Nov 25 16:52:21 crc kubenswrapper[4802]: I1125 16:52:21.545871 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5llz" event={"ID":"b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5","Type":"ContainerStarted","Data":"2ffd8c0afeff16333d67fe9efe2c7ef21f86d94fe8118363190d8a608f96621a"} Nov 25 16:52:21 crc kubenswrapper[4802]: I1125 16:52:21.549558 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlz5g" event={"ID":"ba359221-8966-436b-b846-4c5ac2f1767e","Type":"ContainerStarted","Data":"59625c4c98da7b82b9779e3d22fa4b08bd7d851fca0176f7cc41ba68a430ff8d"} Nov 25 16:52:21 crc kubenswrapper[4802]: I1125 16:52:21.551688 4802 generic.go:334] "Generic (PLEG): container finished" podID="495e0a1d-2dd0-4069-a9dd-6637b835a35a" containerID="44cc7b6a403f668a37564059df525d54004db38ff3bce26a72f0b79e4e6192f0" exitCode=0 Nov 25 16:52:21 crc kubenswrapper[4802]: I1125 16:52:21.551730 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw428" event={"ID":"495e0a1d-2dd0-4069-a9dd-6637b835a35a","Type":"ContainerDied","Data":"44cc7b6a403f668a37564059df525d54004db38ff3bce26a72f0b79e4e6192f0"} Nov 25 16:52:21 crc kubenswrapper[4802]: I1125 16:52:21.551775 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw428" event={"ID":"495e0a1d-2dd0-4069-a9dd-6637b835a35a","Type":"ContainerStarted","Data":"a0a16d9a239bfb7415b85a13b85b5b791b2fbdd4b43576c693e9790ffeec17b1"} Nov 25 16:52:21 crc kubenswrapper[4802]: I1125 16:52:21.584897 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rslrk" podStartSLOduration=2.055881374 podStartE2EDuration="4.584865832s" podCreationTimestamp="2025-11-25 16:52:17 +0000 UTC" firstStartedPulling="2025-11-25 16:52:18.497511344 +0000 UTC m=+321.641858530" lastFinishedPulling="2025-11-25 16:52:21.026495802 +0000 UTC m=+324.170842988" observedRunningTime="2025-11-25 16:52:21.566872687 +0000 UTC m=+324.711220033" watchObservedRunningTime="2025-11-25 16:52:21.584865832 +0000 UTC m=+324.729213018" Nov 25 16:52:21 crc kubenswrapper[4802]: I1125 16:52:21.608658 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mlz5g" podStartSLOduration=3.189836014 podStartE2EDuration="4.608635654s" podCreationTimestamp="2025-11-25 16:52:17 +0000 UTC" firstStartedPulling="2025-11-25 16:52:19.51221803 +0000 UTC m=+322.656565236" lastFinishedPulling="2025-11-25 16:52:20.93101769 +0000 UTC m=+324.075364876" observedRunningTime="2025-11-25 16:52:21.608052507 +0000 UTC m=+324.752399693" watchObservedRunningTime="2025-11-25 16:52:21.608635654 +0000 UTC m=+324.752982840" Nov 25 16:52:22 crc kubenswrapper[4802]: I1125 16:52:22.558798 4802 generic.go:334] "Generic (PLEG): container finished" podID="b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5" containerID="74378f16a5c2088178dc6cab3d4f18bfb3501159d714213d4e4371c403e78a55" exitCode=0 Nov 25 16:52:22 crc kubenswrapper[4802]: I1125 16:52:22.558857 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5llz" event={"ID":"b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5","Type":"ContainerDied","Data":"74378f16a5c2088178dc6cab3d4f18bfb3501159d714213d4e4371c403e78a55"} Nov 25 16:52:24 crc kubenswrapper[4802]: I1125 16:52:24.574330 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5llz" event={"ID":"b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5","Type":"ContainerStarted","Data":"a0050610d3999e604f1e60a524e3b0e19ef797548ebfc61f1d7a073cb6426f79"} Nov 25 16:52:24 crc kubenswrapper[4802]: I1125 16:52:24.576754 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw428" event={"ID":"495e0a1d-2dd0-4069-a9dd-6637b835a35a","Type":"ContainerStarted","Data":"7460a567ac12a1d063c7c8428e24228502a343003a261c2e81da757f09d2bc5f"} Nov 25 16:52:24 crc kubenswrapper[4802]: I1125 16:52:24.599930 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f5llz" podStartSLOduration=3.208964341 podStartE2EDuration="4.599901129s" podCreationTimestamp="2025-11-25 16:52:20 +0000 UTC" firstStartedPulling="2025-11-25 16:52:21.547307807 +0000 UTC m=+324.691654993" lastFinishedPulling="2025-11-25 16:52:22.938244595 +0000 UTC m=+326.082591781" observedRunningTime="2025-11-25 16:52:24.593876055 +0000 UTC m=+327.738223241" watchObservedRunningTime="2025-11-25 16:52:24.599901129 +0000 UTC m=+327.744248315" Nov 25 16:52:25 crc kubenswrapper[4802]: I1125 16:52:25.588976 4802 generic.go:334] "Generic (PLEG): container finished" podID="495e0a1d-2dd0-4069-a9dd-6637b835a35a" containerID="7460a567ac12a1d063c7c8428e24228502a343003a261c2e81da757f09d2bc5f" exitCode=0 Nov 25 16:52:25 crc kubenswrapper[4802]: I1125 16:52:25.589148 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw428" event={"ID":"495e0a1d-2dd0-4069-a9dd-6637b835a35a","Type":"ContainerDied","Data":"7460a567ac12a1d063c7c8428e24228502a343003a261c2e81da757f09d2bc5f"} Nov 25 16:52:26 crc kubenswrapper[4802]: I1125 16:52:26.600603 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw428" event={"ID":"495e0a1d-2dd0-4069-a9dd-6637b835a35a","Type":"ContainerStarted","Data":"3b335721ff28c8b4a895d725611243fadf4a1d1c7ea63a5a45098a973298c9cf"} Nov 25 16:52:26 crc kubenswrapper[4802]: I1125 16:52:26.626506 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kw428" podStartSLOduration=2.160266216 podStartE2EDuration="6.626479506s" podCreationTimestamp="2025-11-25 16:52:20 +0000 UTC" firstStartedPulling="2025-11-25 16:52:21.553370264 +0000 UTC m=+324.697717440" lastFinishedPulling="2025-11-25 16:52:26.019583544 +0000 UTC m=+329.163930730" observedRunningTime="2025-11-25 16:52:26.619283199 +0000 UTC m=+329.763630415" watchObservedRunningTime="2025-11-25 16:52:26.626479506 +0000 UTC m=+329.770826722" Nov 25 16:52:28 crc kubenswrapper[4802]: I1125 16:52:28.033903 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rslrk" Nov 25 16:52:28 crc kubenswrapper[4802]: I1125 16:52:28.034296 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rslrk" Nov 25 16:52:28 crc kubenswrapper[4802]: I1125 16:52:28.072889 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rslrk" Nov 25 16:52:28 crc kubenswrapper[4802]: I1125 16:52:28.220354 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mlz5g" Nov 25 16:52:28 crc kubenswrapper[4802]: I1125 16:52:28.220439 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mlz5g" Nov 25 16:52:28 crc kubenswrapper[4802]: I1125 16:52:28.262101 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mlz5g" Nov 25 16:52:28 crc kubenswrapper[4802]: I1125 16:52:28.654015 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mlz5g" Nov 25 16:52:28 crc kubenswrapper[4802]: I1125 16:52:28.657268 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rslrk" Nov 25 16:52:30 crc kubenswrapper[4802]: I1125 16:52:30.477626 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f5llz" Nov 25 16:52:30 crc kubenswrapper[4802]: I1125 16:52:30.478117 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f5llz" Nov 25 16:52:30 crc kubenswrapper[4802]: I1125 16:52:30.522170 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f5llz" Nov 25 16:52:30 crc kubenswrapper[4802]: I1125 16:52:30.653224 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kw428" Nov 25 16:52:30 crc kubenswrapper[4802]: I1125 16:52:30.653569 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kw428" Nov 25 16:52:30 crc kubenswrapper[4802]: I1125 16:52:30.674546 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f5llz" Nov 25 16:52:31 crc kubenswrapper[4802]: I1125 16:52:31.692721 4802 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kw428" podUID="495e0a1d-2dd0-4069-a9dd-6637b835a35a" containerName="registry-server" probeResult="failure" output=< Nov 25 16:52:31 crc kubenswrapper[4802]: timeout: failed to connect service ":50051" within 1s Nov 25 16:52:31 crc kubenswrapper[4802]: > Nov 25 16:52:40 crc kubenswrapper[4802]: I1125 16:52:40.703664 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kw428" Nov 25 16:52:40 crc kubenswrapper[4802]: I1125 16:52:40.771871 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kw428" Nov 25 16:52:54 crc kubenswrapper[4802]: I1125 16:52:54.248374 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:52:54 crc kubenswrapper[4802]: I1125 16:52:54.249178 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:53:24 crc kubenswrapper[4802]: I1125 16:53:24.248862 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:53:24 crc kubenswrapper[4802]: I1125 16:53:24.249647 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.280491 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wdt8f"] Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.281735 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.305802 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wdt8f"] Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.466311 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/58342705-8111-4ede-8416-7112c44fc758-bound-sa-token\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.466366 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/58342705-8111-4ede-8416-7112c44fc758-registry-tls\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.466402 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/58342705-8111-4ede-8416-7112c44fc758-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.466442 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/58342705-8111-4ede-8416-7112c44fc758-registry-certificates\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.466463 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t26tn\" (UniqueName: \"kubernetes.io/projected/58342705-8111-4ede-8416-7112c44fc758-kube-api-access-t26tn\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.466558 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.466602 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/58342705-8111-4ede-8416-7112c44fc758-trusted-ca\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.466626 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/58342705-8111-4ede-8416-7112c44fc758-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.489541 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.567954 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/58342705-8111-4ede-8416-7112c44fc758-registry-certificates\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.568509 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t26tn\" (UniqueName: \"kubernetes.io/projected/58342705-8111-4ede-8416-7112c44fc758-kube-api-access-t26tn\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.568599 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/58342705-8111-4ede-8416-7112c44fc758-trusted-ca\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.568683 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/58342705-8111-4ede-8416-7112c44fc758-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.569107 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/58342705-8111-4ede-8416-7112c44fc758-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.569341 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/58342705-8111-4ede-8416-7112c44fc758-bound-sa-token\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.569408 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/58342705-8111-4ede-8416-7112c44fc758-registry-tls\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.569487 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/58342705-8111-4ede-8416-7112c44fc758-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.570197 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/58342705-8111-4ede-8416-7112c44fc758-trusted-ca\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.570254 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/58342705-8111-4ede-8416-7112c44fc758-registry-certificates\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.576549 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/58342705-8111-4ede-8416-7112c44fc758-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.576947 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/58342705-8111-4ede-8416-7112c44fc758-registry-tls\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.588719 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t26tn\" (UniqueName: \"kubernetes.io/projected/58342705-8111-4ede-8416-7112c44fc758-kube-api-access-t26tn\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.591455 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/58342705-8111-4ede-8416-7112c44fc758-bound-sa-token\") pod \"image-registry-66df7c8f76-wdt8f\" (UID: \"58342705-8111-4ede-8416-7112c44fc758\") " pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.605694 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.801596 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wdt8f"] Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.989534 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" event={"ID":"58342705-8111-4ede-8416-7112c44fc758","Type":"ContainerStarted","Data":"d4c184646f940440fc3b28a1aa3cde33d84f95f266c8ac29d4f956cbf8002ac5"} Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.990038 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:29 crc kubenswrapper[4802]: I1125 16:53:29.990057 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" event={"ID":"58342705-8111-4ede-8416-7112c44fc758","Type":"ContainerStarted","Data":"ad38613aa99559a5032a949bdb74e269a2c443077c268e2cb84054f03468ef66"} Nov 25 16:53:30 crc kubenswrapper[4802]: I1125 16:53:30.015947 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" podStartSLOduration=1.015922181 podStartE2EDuration="1.015922181s" podCreationTimestamp="2025-11-25 16:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:53:30.011795266 +0000 UTC m=+393.156142442" watchObservedRunningTime="2025-11-25 16:53:30.015922181 +0000 UTC m=+393.160269367" Nov 25 16:53:49 crc kubenswrapper[4802]: I1125 16:53:49.612816 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-wdt8f" Nov 25 16:53:49 crc kubenswrapper[4802]: I1125 16:53:49.668831 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kst7h"] Nov 25 16:53:54 crc kubenswrapper[4802]: I1125 16:53:54.248742 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:53:54 crc kubenswrapper[4802]: I1125 16:53:54.249567 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:53:54 crc kubenswrapper[4802]: I1125 16:53:54.249642 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 16:53:54 crc kubenswrapper[4802]: I1125 16:53:54.250524 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"23620d61a8ef1bbe68d551d6cb4fb38ea02979670932fc175c7a86c5b2d8155b"} pod="openshift-machine-config-operator/machine-config-daemon-h29wc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 16:53:54 crc kubenswrapper[4802]: I1125 16:53:54.250596 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" containerID="cri-o://23620d61a8ef1bbe68d551d6cb4fb38ea02979670932fc175c7a86c5b2d8155b" gracePeriod=600 Nov 25 16:53:55 crc kubenswrapper[4802]: I1125 16:53:55.145853 4802 generic.go:334] "Generic (PLEG): container finished" podID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerID="23620d61a8ef1bbe68d551d6cb4fb38ea02979670932fc175c7a86c5b2d8155b" exitCode=0 Nov 25 16:53:55 crc kubenswrapper[4802]: I1125 16:53:55.145946 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerDied","Data":"23620d61a8ef1bbe68d551d6cb4fb38ea02979670932fc175c7a86c5b2d8155b"} Nov 25 16:53:55 crc kubenswrapper[4802]: I1125 16:53:55.146345 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerStarted","Data":"3d37aca18ca92000cdaea6d946fa2206f34b1d2780852086c38dfe3898b186c8"} Nov 25 16:53:55 crc kubenswrapper[4802]: I1125 16:53:55.146377 4802 scope.go:117] "RemoveContainer" containerID="5bbe1acb407b558db8fb73429c64b4682aab0e148afc2874fa512da66b70faef" Nov 25 16:54:14 crc kubenswrapper[4802]: I1125 16:54:14.710235 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" podUID="8c5d984b-9b31-4997-904a-f27989f78a8c" containerName="registry" containerID="cri-o://247f6204a5cf8a2e11a741690fae81646177dc55c4081145cf72376b1e6e25c3" gracePeriod=30 Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.069971 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.193562 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8c5d984b-9b31-4997-904a-f27989f78a8c-installation-pull-secrets\") pod \"8c5d984b-9b31-4997-904a-f27989f78a8c\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.193674 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8c5d984b-9b31-4997-904a-f27989f78a8c-registry-certificates\") pod \"8c5d984b-9b31-4997-904a-f27989f78a8c\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.193721 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8c5d984b-9b31-4997-904a-f27989f78a8c-ca-trust-extracted\") pod \"8c5d984b-9b31-4997-904a-f27989f78a8c\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.194340 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8c5d984b-9b31-4997-904a-f27989f78a8c\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.194581 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8c5d984b-9b31-4997-904a-f27989f78a8c-registry-tls\") pod \"8c5d984b-9b31-4997-904a-f27989f78a8c\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.194638 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgnl2\" (UniqueName: \"kubernetes.io/projected/8c5d984b-9b31-4997-904a-f27989f78a8c-kube-api-access-dgnl2\") pod \"8c5d984b-9b31-4997-904a-f27989f78a8c\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.194696 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c5d984b-9b31-4997-904a-f27989f78a8c-bound-sa-token\") pod \"8c5d984b-9b31-4997-904a-f27989f78a8c\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.194739 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c5d984b-9b31-4997-904a-f27989f78a8c-trusted-ca\") pod \"8c5d984b-9b31-4997-904a-f27989f78a8c\" (UID: \"8c5d984b-9b31-4997-904a-f27989f78a8c\") " Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.194855 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c5d984b-9b31-4997-904a-f27989f78a8c-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8c5d984b-9b31-4997-904a-f27989f78a8c" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.195460 4802 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8c5d984b-9b31-4997-904a-f27989f78a8c-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.195771 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c5d984b-9b31-4997-904a-f27989f78a8c-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8c5d984b-9b31-4997-904a-f27989f78a8c" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.201713 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c5d984b-9b31-4997-904a-f27989f78a8c-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8c5d984b-9b31-4997-904a-f27989f78a8c" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.201949 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c5d984b-9b31-4997-904a-f27989f78a8c-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8c5d984b-9b31-4997-904a-f27989f78a8c" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.202729 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c5d984b-9b31-4997-904a-f27989f78a8c-kube-api-access-dgnl2" (OuterVolumeSpecName: "kube-api-access-dgnl2") pod "8c5d984b-9b31-4997-904a-f27989f78a8c" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c"). InnerVolumeSpecName "kube-api-access-dgnl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.203229 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c5d984b-9b31-4997-904a-f27989f78a8c-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8c5d984b-9b31-4997-904a-f27989f78a8c" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.208326 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "8c5d984b-9b31-4997-904a-f27989f78a8c" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.218236 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c5d984b-9b31-4997-904a-f27989f78a8c-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8c5d984b-9b31-4997-904a-f27989f78a8c" (UID: "8c5d984b-9b31-4997-904a-f27989f78a8c"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.274345 4802 generic.go:334] "Generic (PLEG): container finished" podID="8c5d984b-9b31-4997-904a-f27989f78a8c" containerID="247f6204a5cf8a2e11a741690fae81646177dc55c4081145cf72376b1e6e25c3" exitCode=0 Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.274406 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.274428 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" event={"ID":"8c5d984b-9b31-4997-904a-f27989f78a8c","Type":"ContainerDied","Data":"247f6204a5cf8a2e11a741690fae81646177dc55c4081145cf72376b1e6e25c3"} Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.274782 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kst7h" event={"ID":"8c5d984b-9b31-4997-904a-f27989f78a8c","Type":"ContainerDied","Data":"0d6e24976433a7ec3a792489e3bf828d6565a5e2d98a176a1cab843c15c9c970"} Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.274803 4802 scope.go:117] "RemoveContainer" containerID="247f6204a5cf8a2e11a741690fae81646177dc55c4081145cf72376b1e6e25c3" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.296452 4802 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8c5d984b-9b31-4997-904a-f27989f78a8c-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.296484 4802 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8c5d984b-9b31-4997-904a-f27989f78a8c-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.296497 4802 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8c5d984b-9b31-4997-904a-f27989f78a8c-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.296510 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgnl2\" (UniqueName: \"kubernetes.io/projected/8c5d984b-9b31-4997-904a-f27989f78a8c-kube-api-access-dgnl2\") on node \"crc\" DevicePath \"\"" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.296520 4802 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c5d984b-9b31-4997-904a-f27989f78a8c-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.296532 4802 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c5d984b-9b31-4997-904a-f27989f78a8c-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.296635 4802 scope.go:117] "RemoveContainer" containerID="247f6204a5cf8a2e11a741690fae81646177dc55c4081145cf72376b1e6e25c3" Nov 25 16:54:15 crc kubenswrapper[4802]: E1125 16:54:15.297166 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"247f6204a5cf8a2e11a741690fae81646177dc55c4081145cf72376b1e6e25c3\": container with ID starting with 247f6204a5cf8a2e11a741690fae81646177dc55c4081145cf72376b1e6e25c3 not found: ID does not exist" containerID="247f6204a5cf8a2e11a741690fae81646177dc55c4081145cf72376b1e6e25c3" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.297213 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"247f6204a5cf8a2e11a741690fae81646177dc55c4081145cf72376b1e6e25c3"} err="failed to get container status \"247f6204a5cf8a2e11a741690fae81646177dc55c4081145cf72376b1e6e25c3\": rpc error: code = NotFound desc = could not find container \"247f6204a5cf8a2e11a741690fae81646177dc55c4081145cf72376b1e6e25c3\": container with ID starting with 247f6204a5cf8a2e11a741690fae81646177dc55c4081145cf72376b1e6e25c3 not found: ID does not exist" Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.306435 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kst7h"] Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.311925 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kst7h"] Nov 25 16:54:15 crc kubenswrapper[4802]: I1125 16:54:15.514563 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c5d984b-9b31-4997-904a-f27989f78a8c" path="/var/lib/kubelet/pods/8c5d984b-9b31-4997-904a-f27989f78a8c/volumes" Nov 25 16:55:54 crc kubenswrapper[4802]: I1125 16:55:54.248761 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:55:54 crc kubenswrapper[4802]: I1125 16:55:54.249919 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:56:24 crc kubenswrapper[4802]: I1125 16:56:24.248634 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:56:24 crc kubenswrapper[4802]: I1125 16:56:24.249806 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:56:54 crc kubenswrapper[4802]: I1125 16:56:54.248678 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:56:54 crc kubenswrapper[4802]: I1125 16:56:54.249470 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:56:54 crc kubenswrapper[4802]: I1125 16:56:54.249546 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 16:56:54 crc kubenswrapper[4802]: I1125 16:56:54.250703 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3d37aca18ca92000cdaea6d946fa2206f34b1d2780852086c38dfe3898b186c8"} pod="openshift-machine-config-operator/machine-config-daemon-h29wc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 16:56:54 crc kubenswrapper[4802]: I1125 16:56:54.250806 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" containerID="cri-o://3d37aca18ca92000cdaea6d946fa2206f34b1d2780852086c38dfe3898b186c8" gracePeriod=600 Nov 25 16:56:55 crc kubenswrapper[4802]: I1125 16:56:55.288726 4802 generic.go:334] "Generic (PLEG): container finished" podID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerID="3d37aca18ca92000cdaea6d946fa2206f34b1d2780852086c38dfe3898b186c8" exitCode=0 Nov 25 16:56:55 crc kubenswrapper[4802]: I1125 16:56:55.288826 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerDied","Data":"3d37aca18ca92000cdaea6d946fa2206f34b1d2780852086c38dfe3898b186c8"} Nov 25 16:56:55 crc kubenswrapper[4802]: I1125 16:56:55.289792 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerStarted","Data":"2dd6a8bbe96b494880e307bbcf47128d8479e3413bfeccced18011d7c6e2daae"} Nov 25 16:56:55 crc kubenswrapper[4802]: I1125 16:56:55.289846 4802 scope.go:117] "RemoveContainer" containerID="23620d61a8ef1bbe68d551d6cb4fb38ea02979670932fc175c7a86c5b2d8155b" Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.637239 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dkxhj"] Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.638523 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovn-controller" containerID="cri-o://90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8" gracePeriod=30 Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.639045 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="sbdb" containerID="cri-o://8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb" gracePeriod=30 Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.639114 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="nbdb" containerID="cri-o://1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4" gracePeriod=30 Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.639213 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="northd" containerID="cri-o://857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542" gracePeriod=30 Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.639271 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816" gracePeriod=30 Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.639324 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="kube-rbac-proxy-node" containerID="cri-o://a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3" gracePeriod=30 Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.639378 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovn-acl-logging" containerID="cri-o://ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8" gracePeriod=30 Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.693053 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovnkube-controller" containerID="cri-o://7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6" gracePeriod=30 Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.843330 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovnkube-controller/3.log" Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.847277 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovn-acl-logging/0.log" Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.847821 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovn-controller/0.log" Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.848159 4802 generic.go:334] "Generic (PLEG): container finished" podID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerID="7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6" exitCode=0 Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.848183 4802 generic.go:334] "Generic (PLEG): container finished" podID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerID="e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816" exitCode=0 Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.848191 4802 generic.go:334] "Generic (PLEG): container finished" podID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerID="a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3" exitCode=0 Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.848198 4802 generic.go:334] "Generic (PLEG): container finished" podID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerID="ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8" exitCode=143 Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.848205 4802 generic.go:334] "Generic (PLEG): container finished" podID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerID="90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8" exitCode=143 Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.848241 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerDied","Data":"7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6"} Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.848272 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerDied","Data":"e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816"} Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.848284 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerDied","Data":"a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3"} Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.848293 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerDied","Data":"ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8"} Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.848302 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerDied","Data":"90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8"} Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.848318 4802 scope.go:117] "RemoveContainer" containerID="88787834da7629acd6521f977ee73dec4231d0b4de6a3bb8f4dad83d203b4221" Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.850739 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hmbvn_97e822d6-58fe-41f4-b08b-3c9b42273307/kube-multus/2.log" Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.851624 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hmbvn_97e822d6-58fe-41f4-b08b-3c9b42273307/kube-multus/1.log" Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.851662 4802 generic.go:334] "Generic (PLEG): container finished" podID="97e822d6-58fe-41f4-b08b-3c9b42273307" containerID="ade8eae2c0adbaed7320d8fd691aabc2c6cf3d3c6b38153dedfccd37ae7a67f8" exitCode=2 Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.851699 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hmbvn" event={"ID":"97e822d6-58fe-41f4-b08b-3c9b42273307","Type":"ContainerDied","Data":"ade8eae2c0adbaed7320d8fd691aabc2c6cf3d3c6b38153dedfccd37ae7a67f8"} Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.852274 4802 scope.go:117] "RemoveContainer" containerID="ade8eae2c0adbaed7320d8fd691aabc2c6cf3d3c6b38153dedfccd37ae7a67f8" Nov 25 16:58:20 crc kubenswrapper[4802]: E1125 16:58:20.852595 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-hmbvn_openshift-multus(97e822d6-58fe-41f4-b08b-3c9b42273307)\"" pod="openshift-multus/multus-hmbvn" podUID="97e822d6-58fe-41f4-b08b-3c9b42273307" Nov 25 16:58:20 crc kubenswrapper[4802]: I1125 16:58:20.945557 4802 scope.go:117] "RemoveContainer" containerID="e0e948183d9361bf5f6831d25601431a1de0662987928a7277917451064c0295" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.008183 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovn-acl-logging/0.log" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.009015 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovn-controller/0.log" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.009524 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.066640 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rk2qk"] Nov 25 16:58:21 crc kubenswrapper[4802]: E1125 16:58:21.066841 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.066854 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 16:58:21 crc kubenswrapper[4802]: E1125 16:58:21.066868 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovnkube-controller" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.066874 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovnkube-controller" Nov 25 16:58:21 crc kubenswrapper[4802]: E1125 16:58:21.066882 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c5d984b-9b31-4997-904a-f27989f78a8c" containerName="registry" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.066888 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c5d984b-9b31-4997-904a-f27989f78a8c" containerName="registry" Nov 25 16:58:21 crc kubenswrapper[4802]: E1125 16:58:21.066895 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="nbdb" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.066901 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="nbdb" Nov 25 16:58:21 crc kubenswrapper[4802]: E1125 16:58:21.066909 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="sbdb" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.066915 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="sbdb" Nov 25 16:58:21 crc kubenswrapper[4802]: E1125 16:58:21.066925 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovnkube-controller" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.066931 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovnkube-controller" Nov 25 16:58:21 crc kubenswrapper[4802]: E1125 16:58:21.066939 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovnkube-controller" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.066945 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovnkube-controller" Nov 25 16:58:21 crc kubenswrapper[4802]: E1125 16:58:21.066951 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="kubecfg-setup" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.066957 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="kubecfg-setup" Nov 25 16:58:21 crc kubenswrapper[4802]: E1125 16:58:21.066965 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="northd" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.066972 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="northd" Nov 25 16:58:21 crc kubenswrapper[4802]: E1125 16:58:21.066978 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovn-acl-logging" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.066984 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovn-acl-logging" Nov 25 16:58:21 crc kubenswrapper[4802]: E1125 16:58:21.066995 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="kube-rbac-proxy-node" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.067001 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="kube-rbac-proxy-node" Nov 25 16:58:21 crc kubenswrapper[4802]: E1125 16:58:21.067009 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovn-controller" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.067015 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovn-controller" Nov 25 16:58:21 crc kubenswrapper[4802]: E1125 16:58:21.067026 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovnkube-controller" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.067032 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovnkube-controller" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.067115 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c5d984b-9b31-4997-904a-f27989f78a8c" containerName="registry" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.067150 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovnkube-controller" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.067157 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovnkube-controller" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.067164 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="kube-rbac-proxy-node" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.067173 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovnkube-controller" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.067183 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="sbdb" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.067207 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="northd" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.067216 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovnkube-controller" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.067225 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.067233 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovn-controller" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.067242 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="nbdb" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.067251 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovn-acl-logging" Nov 25 16:58:21 crc kubenswrapper[4802]: E1125 16:58:21.067330 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovnkube-controller" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.067339 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovnkube-controller" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.067429 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerName="ovnkube-controller" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.068967 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.160618 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-run-netns\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.160753 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-etc-openvswitch\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.160779 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-node-log\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.160806 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.160827 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-log-socket\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.160843 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-cni-netd\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.160864 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-slash\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.160903 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bad5c073-f5d9-4410-9350-bb2a51a764a2-ovnkube-config\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.160919 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-var-lib-openvswitch\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.160956 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bad5c073-f5d9-4410-9350-bb2a51a764a2-env-overrides\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.160989 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-run-openvswitch\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161009 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-cni-bin\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161029 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-run-ovn-kubernetes\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161003 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-log-socket" (OuterVolumeSpecName: "log-socket") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161079 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bad5c073-f5d9-4410-9350-bb2a51a764a2-ovn-node-metrics-cert\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.160988 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-node-log" (OuterVolumeSpecName: "node-log") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161149 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-slash" (OuterVolumeSpecName: "host-slash") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161117 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161093 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161114 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161198 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161222 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161234 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161111 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-systemd-units\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161271 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161303 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161382 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-run-ovn\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161440 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-kubelet\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161365 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161491 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-run-systemd\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161515 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161467 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161593 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bad5c073-f5d9-4410-9350-bb2a51a764a2-ovnkube-script-lib\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161641 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgb5h\" (UniqueName: \"kubernetes.io/projected/bad5c073-f5d9-4410-9350-bb2a51a764a2-kube-api-access-bgb5h\") pod \"bad5c073-f5d9-4410-9350-bb2a51a764a2\" (UID: \"bad5c073-f5d9-4410-9350-bb2a51a764a2\") " Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161707 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bad5c073-f5d9-4410-9350-bb2a51a764a2-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.161828 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bad5c073-f5d9-4410-9350-bb2a51a764a2-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.162167 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bad5c073-f5d9-4410-9350-bb2a51a764a2-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.162416 4802 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.162442 4802 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.162453 4802 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-node-log\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.162466 4802 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.162480 4802 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-log-socket\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.162491 4802 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.162501 4802 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-slash\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.162510 4802 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.162521 4802 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bad5c073-f5d9-4410-9350-bb2a51a764a2-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.162531 4802 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bad5c073-f5d9-4410-9350-bb2a51a764a2-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.162540 4802 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.162549 4802 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.162559 4802 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.162570 4802 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.162579 4802 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.162587 4802 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.162598 4802 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bad5c073-f5d9-4410-9350-bb2a51a764a2-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.168930 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bad5c073-f5d9-4410-9350-bb2a51a764a2-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.169635 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bad5c073-f5d9-4410-9350-bb2a51a764a2-kube-api-access-bgb5h" (OuterVolumeSpecName: "kube-api-access-bgb5h") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "kube-api-access-bgb5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.178734 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "bad5c073-f5d9-4410-9350-bb2a51a764a2" (UID: "bad5c073-f5d9-4410-9350-bb2a51a764a2"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.263667 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-run-ovn\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.263744 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-systemd-units\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.263782 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-var-lib-openvswitch\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.263817 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-run-systemd\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.263876 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.263959 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/913a48c7-b4dc-495f-ba3d-06f7de477ed4-ovnkube-config\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.264007 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vqvs\" (UniqueName: \"kubernetes.io/projected/913a48c7-b4dc-495f-ba3d-06f7de477ed4-kube-api-access-5vqvs\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.264039 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-etc-openvswitch\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.264067 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-log-socket\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.264319 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-kubelet\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.264449 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-slash\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.264497 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-node-log\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.264539 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-cni-bin\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.264588 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-run-ovn-kubernetes\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.264942 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-cni-netd\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.265067 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/913a48c7-b4dc-495f-ba3d-06f7de477ed4-ovnkube-script-lib\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.265229 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/913a48c7-b4dc-495f-ba3d-06f7de477ed4-env-overrides\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.265577 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-run-openvswitch\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.265658 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/913a48c7-b4dc-495f-ba3d-06f7de477ed4-ovn-node-metrics-cert\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.265936 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-run-netns\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.266514 4802 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bad5c073-f5d9-4410-9350-bb2a51a764a2-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.266614 4802 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bad5c073-f5d9-4410-9350-bb2a51a764a2-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.266634 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgb5h\" (UniqueName: \"kubernetes.io/projected/bad5c073-f5d9-4410-9350-bb2a51a764a2-kube-api-access-bgb5h\") on node \"crc\" DevicePath \"\"" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.367565 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.367641 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/913a48c7-b4dc-495f-ba3d-06f7de477ed4-ovnkube-config\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.367666 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vqvs\" (UniqueName: \"kubernetes.io/projected/913a48c7-b4dc-495f-ba3d-06f7de477ed4-kube-api-access-5vqvs\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.367682 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-etc-openvswitch\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.367701 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-log-socket\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.367719 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-kubelet\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.367734 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-slash\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.367750 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-node-log\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.367765 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-cni-bin\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.367785 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-run-ovn-kubernetes\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.367802 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-cni-netd\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.367819 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/913a48c7-b4dc-495f-ba3d-06f7de477ed4-ovnkube-script-lib\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.367815 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-etc-openvswitch\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.367822 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.367860 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-cni-bin\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.367905 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-kubelet\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.367835 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/913a48c7-b4dc-495f-ba3d-06f7de477ed4-env-overrides\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368208 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-log-socket\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368254 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-cni-netd\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368244 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-run-ovn-kubernetes\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368314 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-slash\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368480 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/913a48c7-b4dc-495f-ba3d-06f7de477ed4-env-overrides\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368519 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-node-log\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368639 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-run-openvswitch\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368670 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/913a48c7-b4dc-495f-ba3d-06f7de477ed4-ovn-node-metrics-cert\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368695 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-run-netns\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368723 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-systemd-units\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368743 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-run-ovn\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368765 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-var-lib-openvswitch\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368786 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-run-systemd\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368780 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-run-openvswitch\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368867 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-run-ovn\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368873 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-systemd-units\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368900 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-var-lib-openvswitch\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368933 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-run-systemd\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.368960 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/913a48c7-b4dc-495f-ba3d-06f7de477ed4-host-run-netns\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.369410 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/913a48c7-b4dc-495f-ba3d-06f7de477ed4-ovnkube-config\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.369470 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/913a48c7-b4dc-495f-ba3d-06f7de477ed4-ovnkube-script-lib\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.375796 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/913a48c7-b4dc-495f-ba3d-06f7de477ed4-ovn-node-metrics-cert\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.396061 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vqvs\" (UniqueName: \"kubernetes.io/projected/913a48c7-b4dc-495f-ba3d-06f7de477ed4-kube-api-access-5vqvs\") pod \"ovnkube-node-rk2qk\" (UID: \"913a48c7-b4dc-495f-ba3d-06f7de477ed4\") " pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.687503 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.861532 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovn-acl-logging/0.log" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.862104 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dkxhj_bad5c073-f5d9-4410-9350-bb2a51a764a2/ovn-controller/0.log" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.862915 4802 generic.go:334] "Generic (PLEG): container finished" podID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerID="8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb" exitCode=0 Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.862943 4802 generic.go:334] "Generic (PLEG): container finished" podID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerID="1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4" exitCode=0 Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.862950 4802 generic.go:334] "Generic (PLEG): container finished" podID="bad5c073-f5d9-4410-9350-bb2a51a764a2" containerID="857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542" exitCode=0 Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.863038 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.863041 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerDied","Data":"8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb"} Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.863136 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerDied","Data":"1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4"} Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.863160 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerDied","Data":"857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542"} Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.863164 4802 scope.go:117] "RemoveContainer" containerID="7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.863179 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dkxhj" event={"ID":"bad5c073-f5d9-4410-9350-bb2a51a764a2","Type":"ContainerDied","Data":"eaf11c74cfd2e012d33899947df8e98203e35baa2ed68906bf688745fbeaed6c"} Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.866177 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" event={"ID":"913a48c7-b4dc-495f-ba3d-06f7de477ed4","Type":"ContainerStarted","Data":"7bf4bc4c4f6c43d8da7e54e76a149d098dc5e51f6e9c105539181d92ad29d0d9"} Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.868396 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hmbvn_97e822d6-58fe-41f4-b08b-3c9b42273307/kube-multus/2.log" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.912655 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dkxhj"] Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.916723 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dkxhj"] Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.926528 4802 scope.go:117] "RemoveContainer" containerID="8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.938708 4802 scope.go:117] "RemoveContainer" containerID="1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.949357 4802 scope.go:117] "RemoveContainer" containerID="857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.961819 4802 scope.go:117] "RemoveContainer" containerID="e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.975321 4802 scope.go:117] "RemoveContainer" containerID="a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3" Nov 25 16:58:21 crc kubenswrapper[4802]: I1125 16:58:21.988811 4802 scope.go:117] "RemoveContainer" containerID="ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.000328 4802 scope.go:117] "RemoveContainer" containerID="90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.014182 4802 scope.go:117] "RemoveContainer" containerID="c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.035951 4802 scope.go:117] "RemoveContainer" containerID="7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6" Nov 25 16:58:22 crc kubenswrapper[4802]: E1125 16:58:22.036569 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6\": container with ID starting with 7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6 not found: ID does not exist" containerID="7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.036627 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6"} err="failed to get container status \"7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6\": rpc error: code = NotFound desc = could not find container \"7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6\": container with ID starting with 7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.036664 4802 scope.go:117] "RemoveContainer" containerID="8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb" Nov 25 16:58:22 crc kubenswrapper[4802]: E1125 16:58:22.037077 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\": container with ID starting with 8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb not found: ID does not exist" containerID="8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.037109 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb"} err="failed to get container status \"8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\": rpc error: code = NotFound desc = could not find container \"8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\": container with ID starting with 8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.037142 4802 scope.go:117] "RemoveContainer" containerID="1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4" Nov 25 16:58:22 crc kubenswrapper[4802]: E1125 16:58:22.037674 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\": container with ID starting with 1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4 not found: ID does not exist" containerID="1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.037724 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4"} err="failed to get container status \"1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\": rpc error: code = NotFound desc = could not find container \"1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\": container with ID starting with 1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.037757 4802 scope.go:117] "RemoveContainer" containerID="857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542" Nov 25 16:58:22 crc kubenswrapper[4802]: E1125 16:58:22.038247 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\": container with ID starting with 857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542 not found: ID does not exist" containerID="857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.038314 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542"} err="failed to get container status \"857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\": rpc error: code = NotFound desc = could not find container \"857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\": container with ID starting with 857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.038356 4802 scope.go:117] "RemoveContainer" containerID="e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816" Nov 25 16:58:22 crc kubenswrapper[4802]: E1125 16:58:22.038728 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\": container with ID starting with e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816 not found: ID does not exist" containerID="e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.038764 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816"} err="failed to get container status \"e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\": rpc error: code = NotFound desc = could not find container \"e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\": container with ID starting with e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.038783 4802 scope.go:117] "RemoveContainer" containerID="a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3" Nov 25 16:58:22 crc kubenswrapper[4802]: E1125 16:58:22.039040 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\": container with ID starting with a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3 not found: ID does not exist" containerID="a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.039077 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3"} err="failed to get container status \"a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\": rpc error: code = NotFound desc = could not find container \"a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\": container with ID starting with a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.039099 4802 scope.go:117] "RemoveContainer" containerID="ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8" Nov 25 16:58:22 crc kubenswrapper[4802]: E1125 16:58:22.039351 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\": container with ID starting with ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8 not found: ID does not exist" containerID="ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.039379 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8"} err="failed to get container status \"ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\": rpc error: code = NotFound desc = could not find container \"ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\": container with ID starting with ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.039398 4802 scope.go:117] "RemoveContainer" containerID="90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8" Nov 25 16:58:22 crc kubenswrapper[4802]: E1125 16:58:22.039643 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\": container with ID starting with 90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8 not found: ID does not exist" containerID="90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.039776 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8"} err="failed to get container status \"90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\": rpc error: code = NotFound desc = could not find container \"90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\": container with ID starting with 90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.039804 4802 scope.go:117] "RemoveContainer" containerID="c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99" Nov 25 16:58:22 crc kubenswrapper[4802]: E1125 16:58:22.040045 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\": container with ID starting with c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99 not found: ID does not exist" containerID="c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.040071 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99"} err="failed to get container status \"c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\": rpc error: code = NotFound desc = could not find container \"c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\": container with ID starting with c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.040092 4802 scope.go:117] "RemoveContainer" containerID="7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.040427 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6"} err="failed to get container status \"7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6\": rpc error: code = NotFound desc = could not find container \"7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6\": container with ID starting with 7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.040466 4802 scope.go:117] "RemoveContainer" containerID="8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.040914 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb"} err="failed to get container status \"8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\": rpc error: code = NotFound desc = could not find container \"8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\": container with ID starting with 8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.040937 4802 scope.go:117] "RemoveContainer" containerID="1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.041201 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4"} err="failed to get container status \"1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\": rpc error: code = NotFound desc = could not find container \"1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\": container with ID starting with 1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.041238 4802 scope.go:117] "RemoveContainer" containerID="857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.041485 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542"} err="failed to get container status \"857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\": rpc error: code = NotFound desc = could not find container \"857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\": container with ID starting with 857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.041522 4802 scope.go:117] "RemoveContainer" containerID="e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.041781 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816"} err="failed to get container status \"e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\": rpc error: code = NotFound desc = could not find container \"e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\": container with ID starting with e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.041811 4802 scope.go:117] "RemoveContainer" containerID="a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.042045 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3"} err="failed to get container status \"a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\": rpc error: code = NotFound desc = could not find container \"a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\": container with ID starting with a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.042074 4802 scope.go:117] "RemoveContainer" containerID="ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.042312 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8"} err="failed to get container status \"ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\": rpc error: code = NotFound desc = could not find container \"ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\": container with ID starting with ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.042333 4802 scope.go:117] "RemoveContainer" containerID="90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.042563 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8"} err="failed to get container status \"90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\": rpc error: code = NotFound desc = could not find container \"90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\": container with ID starting with 90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.042595 4802 scope.go:117] "RemoveContainer" containerID="c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.042884 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99"} err="failed to get container status \"c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\": rpc error: code = NotFound desc = could not find container \"c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\": container with ID starting with c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.042905 4802 scope.go:117] "RemoveContainer" containerID="7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.043150 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6"} err="failed to get container status \"7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6\": rpc error: code = NotFound desc = could not find container \"7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6\": container with ID starting with 7f39382410ce7270bf609c86a56d973d6125836c6c88d687e30f76dd420b86d6 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.043182 4802 scope.go:117] "RemoveContainer" containerID="8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.043433 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb"} err="failed to get container status \"8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\": rpc error: code = NotFound desc = could not find container \"8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb\": container with ID starting with 8004b9c762ca051f1b4fe742762507469e18183756933a7b90856a54daac94fb not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.043501 4802 scope.go:117] "RemoveContainer" containerID="1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.043728 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4"} err="failed to get container status \"1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\": rpc error: code = NotFound desc = could not find container \"1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4\": container with ID starting with 1113b20b3e3231c640b694546a748708424f7861326c76a819b6880b3073bbc4 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.043749 4802 scope.go:117] "RemoveContainer" containerID="857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.043956 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542"} err="failed to get container status \"857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\": rpc error: code = NotFound desc = could not find container \"857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542\": container with ID starting with 857a9ffd103b610d15fa2b8b6f3c56f529d20fc3e5b05b4e02dad677893a2542 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.043980 4802 scope.go:117] "RemoveContainer" containerID="e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.044226 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816"} err="failed to get container status \"e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\": rpc error: code = NotFound desc = could not find container \"e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816\": container with ID starting with e5e763037eb29ba76fa1fd7fbb57f21c6fd2c3eaae5e2893a1df9a6bf0c2c816 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.044244 4802 scope.go:117] "RemoveContainer" containerID="a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.044457 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3"} err="failed to get container status \"a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\": rpc error: code = NotFound desc = could not find container \"a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3\": container with ID starting with a918219e9bd50c36740c3d84d27dbb4faaa143a4f5abbb61061e0d5bdafdb3b3 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.044488 4802 scope.go:117] "RemoveContainer" containerID="ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.044906 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8"} err="failed to get container status \"ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\": rpc error: code = NotFound desc = could not find container \"ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8\": container with ID starting with ba590ca2f1b55f07cd71501c7807e12a6a0b0d294baaf8b6732f643aeaf92ce8 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.044939 4802 scope.go:117] "RemoveContainer" containerID="90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.045193 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8"} err="failed to get container status \"90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\": rpc error: code = NotFound desc = could not find container \"90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8\": container with ID starting with 90ee43ea0bd3fbae0750e80c0fc8eeb8e1e6b89f6f20a1c60a6dc197869783d8 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.045227 4802 scope.go:117] "RemoveContainer" containerID="c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.045472 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99"} err="failed to get container status \"c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\": rpc error: code = NotFound desc = could not find container \"c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99\": container with ID starting with c9ad9b905c4cea927ca596de99d825e08f452ee887c553c6a510cbaad6f0fe99 not found: ID does not exist" Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.880050 4802 generic.go:334] "Generic (PLEG): container finished" podID="913a48c7-b4dc-495f-ba3d-06f7de477ed4" containerID="0350c6fdb49b433b3bb7aa1cef7267fbab7bdc12637103ba990c5cf4cc362810" exitCode=0 Nov 25 16:58:22 crc kubenswrapper[4802]: I1125 16:58:22.880177 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" event={"ID":"913a48c7-b4dc-495f-ba3d-06f7de477ed4","Type":"ContainerDied","Data":"0350c6fdb49b433b3bb7aa1cef7267fbab7bdc12637103ba990c5cf4cc362810"} Nov 25 16:58:23 crc kubenswrapper[4802]: I1125 16:58:23.514224 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bad5c073-f5d9-4410-9350-bb2a51a764a2" path="/var/lib/kubelet/pods/bad5c073-f5d9-4410-9350-bb2a51a764a2/volumes" Nov 25 16:58:23 crc kubenswrapper[4802]: I1125 16:58:23.890027 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" event={"ID":"913a48c7-b4dc-495f-ba3d-06f7de477ed4","Type":"ContainerStarted","Data":"ff25b61b8c39978458ce09e3ec4ef10a6cddd5b1d1528d598ca1057ee45af632"} Nov 25 16:58:23 crc kubenswrapper[4802]: I1125 16:58:23.890468 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" event={"ID":"913a48c7-b4dc-495f-ba3d-06f7de477ed4","Type":"ContainerStarted","Data":"ec3dd6d9c541a2b073ed8abed5abcdcfaa4274cb23ead018f84845ab72862bc3"} Nov 25 16:58:23 crc kubenswrapper[4802]: I1125 16:58:23.890480 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" event={"ID":"913a48c7-b4dc-495f-ba3d-06f7de477ed4","Type":"ContainerStarted","Data":"c1a5c254d2938cdf65bfb9be08ed78e84e6124995739c080b77af71284b995f1"} Nov 25 16:58:23 crc kubenswrapper[4802]: I1125 16:58:23.890489 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" event={"ID":"913a48c7-b4dc-495f-ba3d-06f7de477ed4","Type":"ContainerStarted","Data":"9218c1149832b0dac352a8c5dc76f3f384fde1ec000dd019f5b6238659353907"} Nov 25 16:58:23 crc kubenswrapper[4802]: I1125 16:58:23.890497 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" event={"ID":"913a48c7-b4dc-495f-ba3d-06f7de477ed4","Type":"ContainerStarted","Data":"a3b12eeb00c9a7914ba2276a6a40e1eb587de13bcc52a3bcbc98a91e7356784b"} Nov 25 16:58:23 crc kubenswrapper[4802]: I1125 16:58:23.890507 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" event={"ID":"913a48c7-b4dc-495f-ba3d-06f7de477ed4","Type":"ContainerStarted","Data":"d0a6dc64468b20548a30680abd3cc45397d2ce379f7f6f0f366a1930f8b7912c"} Nov 25 16:58:25 crc kubenswrapper[4802]: I1125 16:58:25.908241 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" event={"ID":"913a48c7-b4dc-495f-ba3d-06f7de477ed4","Type":"ContainerStarted","Data":"51bcbe30a927560b11497be70f776c6547504fc3add7f95c9707c42242acba5c"} Nov 25 16:58:28 crc kubenswrapper[4802]: I1125 16:58:28.933462 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" event={"ID":"913a48c7-b4dc-495f-ba3d-06f7de477ed4","Type":"ContainerStarted","Data":"edeefe37a04a513a96612947b131ff390c25fa45bac619fb17ddc49d1a3ec77b"} Nov 25 16:58:28 crc kubenswrapper[4802]: I1125 16:58:28.934511 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:28 crc kubenswrapper[4802]: I1125 16:58:28.934530 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:28 crc kubenswrapper[4802]: I1125 16:58:28.970685 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" podStartSLOduration=7.970659315 podStartE2EDuration="7.970659315s" podCreationTimestamp="2025-11-25 16:58:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:58:28.964304753 +0000 UTC m=+692.108651959" watchObservedRunningTime="2025-11-25 16:58:28.970659315 +0000 UTC m=+692.115006501" Nov 25 16:58:28 crc kubenswrapper[4802]: I1125 16:58:28.979819 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:29 crc kubenswrapper[4802]: I1125 16:58:29.941355 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:29 crc kubenswrapper[4802]: I1125 16:58:29.988488 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:32 crc kubenswrapper[4802]: I1125 16:58:32.659887 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4"] Nov 25 16:58:32 crc kubenswrapper[4802]: I1125 16:58:32.661501 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:32 crc kubenswrapper[4802]: I1125 16:58:32.664325 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 16:58:32 crc kubenswrapper[4802]: I1125 16:58:32.677755 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4"] Nov 25 16:58:32 crc kubenswrapper[4802]: I1125 16:58:32.731269 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ef5d76c-957b-4785-8165-50d2b55790c7-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4\" (UID: \"2ef5d76c-957b-4785-8165-50d2b55790c7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:32 crc kubenswrapper[4802]: I1125 16:58:32.731438 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mss2b\" (UniqueName: \"kubernetes.io/projected/2ef5d76c-957b-4785-8165-50d2b55790c7-kube-api-access-mss2b\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4\" (UID: \"2ef5d76c-957b-4785-8165-50d2b55790c7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:32 crc kubenswrapper[4802]: I1125 16:58:32.731490 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ef5d76c-957b-4785-8165-50d2b55790c7-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4\" (UID: \"2ef5d76c-957b-4785-8165-50d2b55790c7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:32 crc kubenswrapper[4802]: I1125 16:58:32.832252 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mss2b\" (UniqueName: \"kubernetes.io/projected/2ef5d76c-957b-4785-8165-50d2b55790c7-kube-api-access-mss2b\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4\" (UID: \"2ef5d76c-957b-4785-8165-50d2b55790c7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:32 crc kubenswrapper[4802]: I1125 16:58:32.832328 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ef5d76c-957b-4785-8165-50d2b55790c7-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4\" (UID: \"2ef5d76c-957b-4785-8165-50d2b55790c7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:32 crc kubenswrapper[4802]: I1125 16:58:32.832410 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ef5d76c-957b-4785-8165-50d2b55790c7-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4\" (UID: \"2ef5d76c-957b-4785-8165-50d2b55790c7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:32 crc kubenswrapper[4802]: I1125 16:58:32.833015 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ef5d76c-957b-4785-8165-50d2b55790c7-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4\" (UID: \"2ef5d76c-957b-4785-8165-50d2b55790c7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:32 crc kubenswrapper[4802]: I1125 16:58:32.833089 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ef5d76c-957b-4785-8165-50d2b55790c7-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4\" (UID: \"2ef5d76c-957b-4785-8165-50d2b55790c7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:32 crc kubenswrapper[4802]: I1125 16:58:32.856085 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mss2b\" (UniqueName: \"kubernetes.io/projected/2ef5d76c-957b-4785-8165-50d2b55790c7-kube-api-access-mss2b\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4\" (UID: \"2ef5d76c-957b-4785-8165-50d2b55790c7\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:32 crc kubenswrapper[4802]: I1125 16:58:32.994041 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:33 crc kubenswrapper[4802]: E1125 16:58:33.021149 4802 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_openshift-marketplace_2ef5d76c-957b-4785-8165-50d2b55790c7_0(c614437ce74051f85d3264805d69e50d435447ff223647f4c5dbcff36c4d8018): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 16:58:33 crc kubenswrapper[4802]: E1125 16:58:33.021241 4802 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_openshift-marketplace_2ef5d76c-957b-4785-8165-50d2b55790c7_0(c614437ce74051f85d3264805d69e50d435447ff223647f4c5dbcff36c4d8018): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:33 crc kubenswrapper[4802]: E1125 16:58:33.021275 4802 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_openshift-marketplace_2ef5d76c-957b-4785-8165-50d2b55790c7_0(c614437ce74051f85d3264805d69e50d435447ff223647f4c5dbcff36c4d8018): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:33 crc kubenswrapper[4802]: E1125 16:58:33.021341 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_openshift-marketplace(2ef5d76c-957b-4785-8165-50d2b55790c7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_openshift-marketplace(2ef5d76c-957b-4785-8165-50d2b55790c7)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_openshift-marketplace_2ef5d76c-957b-4785-8165-50d2b55790c7_0(c614437ce74051f85d3264805d69e50d435447ff223647f4c5dbcff36c4d8018): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" podUID="2ef5d76c-957b-4785-8165-50d2b55790c7" Nov 25 16:58:33 crc kubenswrapper[4802]: I1125 16:58:33.962756 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:33 crc kubenswrapper[4802]: I1125 16:58:33.963295 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:33 crc kubenswrapper[4802]: E1125 16:58:33.990562 4802 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_openshift-marketplace_2ef5d76c-957b-4785-8165-50d2b55790c7_0(0e12efc7ed297d1db819975a2915e0edbcbd1d5d7c0fd53ac47d1d8d789a84cb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 16:58:33 crc kubenswrapper[4802]: E1125 16:58:33.990692 4802 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_openshift-marketplace_2ef5d76c-957b-4785-8165-50d2b55790c7_0(0e12efc7ed297d1db819975a2915e0edbcbd1d5d7c0fd53ac47d1d8d789a84cb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:33 crc kubenswrapper[4802]: E1125 16:58:33.990738 4802 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_openshift-marketplace_2ef5d76c-957b-4785-8165-50d2b55790c7_0(0e12efc7ed297d1db819975a2915e0edbcbd1d5d7c0fd53ac47d1d8d789a84cb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:33 crc kubenswrapper[4802]: E1125 16:58:33.990826 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_openshift-marketplace(2ef5d76c-957b-4785-8165-50d2b55790c7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_openshift-marketplace(2ef5d76c-957b-4785-8165-50d2b55790c7)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_openshift-marketplace_2ef5d76c-957b-4785-8165-50d2b55790c7_0(0e12efc7ed297d1db819975a2915e0edbcbd1d5d7c0fd53ac47d1d8d789a84cb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" podUID="2ef5d76c-957b-4785-8165-50d2b55790c7" Nov 25 16:58:34 crc kubenswrapper[4802]: I1125 16:58:34.504965 4802 scope.go:117] "RemoveContainer" containerID="ade8eae2c0adbaed7320d8fd691aabc2c6cf3d3c6b38153dedfccd37ae7a67f8" Nov 25 16:58:34 crc kubenswrapper[4802]: E1125 16:58:34.505448 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-hmbvn_openshift-multus(97e822d6-58fe-41f4-b08b-3c9b42273307)\"" pod="openshift-multus/multus-hmbvn" podUID="97e822d6-58fe-41f4-b08b-3c9b42273307" Nov 25 16:58:46 crc kubenswrapper[4802]: I1125 16:58:46.504256 4802 scope.go:117] "RemoveContainer" containerID="ade8eae2c0adbaed7320d8fd691aabc2c6cf3d3c6b38153dedfccd37ae7a67f8" Nov 25 16:58:47 crc kubenswrapper[4802]: I1125 16:58:47.061212 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hmbvn_97e822d6-58fe-41f4-b08b-3c9b42273307/kube-multus/2.log" Nov 25 16:58:47 crc kubenswrapper[4802]: I1125 16:58:47.061867 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hmbvn" event={"ID":"97e822d6-58fe-41f4-b08b-3c9b42273307","Type":"ContainerStarted","Data":"1c4dd58be2bdfd4de567b571552cb7a77ccbd93f8725264909ff674b77e9804e"} Nov 25 16:58:47 crc kubenswrapper[4802]: I1125 16:58:47.503897 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:47 crc kubenswrapper[4802]: I1125 16:58:47.509622 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:47 crc kubenswrapper[4802]: E1125 16:58:47.553076 4802 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_openshift-marketplace_2ef5d76c-957b-4785-8165-50d2b55790c7_0(732aa3f88e2beb20afd59867c870782da7e496facb966333845d0628b826618b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 16:58:47 crc kubenswrapper[4802]: E1125 16:58:47.553341 4802 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_openshift-marketplace_2ef5d76c-957b-4785-8165-50d2b55790c7_0(732aa3f88e2beb20afd59867c870782da7e496facb966333845d0628b826618b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:47 crc kubenswrapper[4802]: E1125 16:58:47.553455 4802 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_openshift-marketplace_2ef5d76c-957b-4785-8165-50d2b55790c7_0(732aa3f88e2beb20afd59867c870782da7e496facb966333845d0628b826618b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:47 crc kubenswrapper[4802]: E1125 16:58:47.553597 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_openshift-marketplace(2ef5d76c-957b-4785-8165-50d2b55790c7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_openshift-marketplace(2ef5d76c-957b-4785-8165-50d2b55790c7)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_openshift-marketplace_2ef5d76c-957b-4785-8165-50d2b55790c7_0(732aa3f88e2beb20afd59867c870782da7e496facb966333845d0628b826618b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" podUID="2ef5d76c-957b-4785-8165-50d2b55790c7" Nov 25 16:58:51 crc kubenswrapper[4802]: I1125 16:58:51.711468 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rk2qk" Nov 25 16:58:54 crc kubenswrapper[4802]: I1125 16:58:54.249386 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:58:54 crc kubenswrapper[4802]: I1125 16:58:54.249515 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:58:59 crc kubenswrapper[4802]: I1125 16:58:59.504419 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:59 crc kubenswrapper[4802]: I1125 16:58:59.505810 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:58:59 crc kubenswrapper[4802]: I1125 16:58:59.695997 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4"] Nov 25 16:58:59 crc kubenswrapper[4802]: W1125 16:58:59.705200 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ef5d76c_957b_4785_8165_50d2b55790c7.slice/crio-61ab7422dd0bd7a376f74eb62aaf25617086c043fa53f1498f9c4e15648564d2 WatchSource:0}: Error finding container 61ab7422dd0bd7a376f74eb62aaf25617086c043fa53f1498f9c4e15648564d2: Status 404 returned error can't find the container with id 61ab7422dd0bd7a376f74eb62aaf25617086c043fa53f1498f9c4e15648564d2 Nov 25 16:59:00 crc kubenswrapper[4802]: I1125 16:59:00.141910 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" event={"ID":"2ef5d76c-957b-4785-8165-50d2b55790c7","Type":"ContainerStarted","Data":"156918bf1b1d42b4628ae384bf5f0eeca11ee1d7c2f5c4bd1e66bb89d4c031fa"} Nov 25 16:59:00 crc kubenswrapper[4802]: I1125 16:59:00.141991 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" event={"ID":"2ef5d76c-957b-4785-8165-50d2b55790c7","Type":"ContainerStarted","Data":"61ab7422dd0bd7a376f74eb62aaf25617086c043fa53f1498f9c4e15648564d2"} Nov 25 16:59:01 crc kubenswrapper[4802]: I1125 16:59:01.150618 4802 generic.go:334] "Generic (PLEG): container finished" podID="2ef5d76c-957b-4785-8165-50d2b55790c7" containerID="156918bf1b1d42b4628ae384bf5f0eeca11ee1d7c2f5c4bd1e66bb89d4c031fa" exitCode=0 Nov 25 16:59:01 crc kubenswrapper[4802]: I1125 16:59:01.150719 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" event={"ID":"2ef5d76c-957b-4785-8165-50d2b55790c7","Type":"ContainerDied","Data":"156918bf1b1d42b4628ae384bf5f0eeca11ee1d7c2f5c4bd1e66bb89d4c031fa"} Nov 25 16:59:01 crc kubenswrapper[4802]: I1125 16:59:01.153582 4802 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 16:59:03 crc kubenswrapper[4802]: I1125 16:59:03.168876 4802 generic.go:334] "Generic (PLEG): container finished" podID="2ef5d76c-957b-4785-8165-50d2b55790c7" containerID="64d275f61067d81538c4836ea7e4ceceb7e440edca1ba336d31f10f6ec89aa70" exitCode=0 Nov 25 16:59:03 crc kubenswrapper[4802]: I1125 16:59:03.169345 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" event={"ID":"2ef5d76c-957b-4785-8165-50d2b55790c7","Type":"ContainerDied","Data":"64d275f61067d81538c4836ea7e4ceceb7e440edca1ba336d31f10f6ec89aa70"} Nov 25 16:59:04 crc kubenswrapper[4802]: I1125 16:59:04.176788 4802 generic.go:334] "Generic (PLEG): container finished" podID="2ef5d76c-957b-4785-8165-50d2b55790c7" containerID="3865bcb46a7de1d001a05c08b5cec738a7806731d10414134ed6839e72ad9c90" exitCode=0 Nov 25 16:59:04 crc kubenswrapper[4802]: I1125 16:59:04.176840 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" event={"ID":"2ef5d76c-957b-4785-8165-50d2b55790c7","Type":"ContainerDied","Data":"3865bcb46a7de1d001a05c08b5cec738a7806731d10414134ed6839e72ad9c90"} Nov 25 16:59:05 crc kubenswrapper[4802]: I1125 16:59:05.409978 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:59:05 crc kubenswrapper[4802]: I1125 16:59:05.516864 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ef5d76c-957b-4785-8165-50d2b55790c7-util\") pod \"2ef5d76c-957b-4785-8165-50d2b55790c7\" (UID: \"2ef5d76c-957b-4785-8165-50d2b55790c7\") " Nov 25 16:59:05 crc kubenswrapper[4802]: I1125 16:59:05.517037 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mss2b\" (UniqueName: \"kubernetes.io/projected/2ef5d76c-957b-4785-8165-50d2b55790c7-kube-api-access-mss2b\") pod \"2ef5d76c-957b-4785-8165-50d2b55790c7\" (UID: \"2ef5d76c-957b-4785-8165-50d2b55790c7\") " Nov 25 16:59:05 crc kubenswrapper[4802]: I1125 16:59:05.517113 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ef5d76c-957b-4785-8165-50d2b55790c7-bundle\") pod \"2ef5d76c-957b-4785-8165-50d2b55790c7\" (UID: \"2ef5d76c-957b-4785-8165-50d2b55790c7\") " Nov 25 16:59:05 crc kubenswrapper[4802]: I1125 16:59:05.519032 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ef5d76c-957b-4785-8165-50d2b55790c7-bundle" (OuterVolumeSpecName: "bundle") pod "2ef5d76c-957b-4785-8165-50d2b55790c7" (UID: "2ef5d76c-957b-4785-8165-50d2b55790c7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:59:05 crc kubenswrapper[4802]: I1125 16:59:05.524060 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ef5d76c-957b-4785-8165-50d2b55790c7-kube-api-access-mss2b" (OuterVolumeSpecName: "kube-api-access-mss2b") pod "2ef5d76c-957b-4785-8165-50d2b55790c7" (UID: "2ef5d76c-957b-4785-8165-50d2b55790c7"). InnerVolumeSpecName "kube-api-access-mss2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:59:05 crc kubenswrapper[4802]: I1125 16:59:05.619695 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mss2b\" (UniqueName: \"kubernetes.io/projected/2ef5d76c-957b-4785-8165-50d2b55790c7-kube-api-access-mss2b\") on node \"crc\" DevicePath \"\"" Nov 25 16:59:05 crc kubenswrapper[4802]: I1125 16:59:05.620105 4802 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ef5d76c-957b-4785-8165-50d2b55790c7-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 16:59:05 crc kubenswrapper[4802]: I1125 16:59:05.742746 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ef5d76c-957b-4785-8165-50d2b55790c7-util" (OuterVolumeSpecName: "util") pod "2ef5d76c-957b-4785-8165-50d2b55790c7" (UID: "2ef5d76c-957b-4785-8165-50d2b55790c7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:59:05 crc kubenswrapper[4802]: I1125 16:59:05.823065 4802 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ef5d76c-957b-4785-8165-50d2b55790c7-util\") on node \"crc\" DevicePath \"\"" Nov 25 16:59:06 crc kubenswrapper[4802]: I1125 16:59:06.191428 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" event={"ID":"2ef5d76c-957b-4785-8165-50d2b55790c7","Type":"ContainerDied","Data":"61ab7422dd0bd7a376f74eb62aaf25617086c043fa53f1498f9c4e15648564d2"} Nov 25 16:59:06 crc kubenswrapper[4802]: I1125 16:59:06.191488 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61ab7422dd0bd7a376f74eb62aaf25617086c043fa53f1498f9c4e15648564d2" Nov 25 16:59:06 crc kubenswrapper[4802]: I1125 16:59:06.191539 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4" Nov 25 16:59:15 crc kubenswrapper[4802]: I1125 16:59:15.962056 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f"] Nov 25 16:59:15 crc kubenswrapper[4802]: E1125 16:59:15.965080 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ef5d76c-957b-4785-8165-50d2b55790c7" containerName="util" Nov 25 16:59:15 crc kubenswrapper[4802]: I1125 16:59:15.965099 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ef5d76c-957b-4785-8165-50d2b55790c7" containerName="util" Nov 25 16:59:15 crc kubenswrapper[4802]: E1125 16:59:15.965109 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ef5d76c-957b-4785-8165-50d2b55790c7" containerName="pull" Nov 25 16:59:15 crc kubenswrapper[4802]: I1125 16:59:15.965115 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ef5d76c-957b-4785-8165-50d2b55790c7" containerName="pull" Nov 25 16:59:15 crc kubenswrapper[4802]: E1125 16:59:15.965153 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ef5d76c-957b-4785-8165-50d2b55790c7" containerName="extract" Nov 25 16:59:15 crc kubenswrapper[4802]: I1125 16:59:15.965159 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ef5d76c-957b-4785-8165-50d2b55790c7" containerName="extract" Nov 25 16:59:15 crc kubenswrapper[4802]: I1125 16:59:15.965390 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ef5d76c-957b-4785-8165-50d2b55790c7" containerName="extract" Nov 25 16:59:15 crc kubenswrapper[4802]: I1125 16:59:15.965967 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" Nov 25 16:59:15 crc kubenswrapper[4802]: I1125 16:59:15.976336 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 25 16:59:15 crc kubenswrapper[4802]: I1125 16:59:15.976501 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 25 16:59:15 crc kubenswrapper[4802]: I1125 16:59:15.976336 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 25 16:59:15 crc kubenswrapper[4802]: I1125 16:59:15.976727 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-q2wcj" Nov 25 16:59:15 crc kubenswrapper[4802]: I1125 16:59:15.976988 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 25 16:59:15 crc kubenswrapper[4802]: I1125 16:59:15.987785 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f"] Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.057802 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4j89\" (UniqueName: \"kubernetes.io/projected/a1724655-9ac1-43dc-8292-f39870b4a855-kube-api-access-g4j89\") pod \"metallb-operator-controller-manager-548f76d7c-s5g2f\" (UID: \"a1724655-9ac1-43dc-8292-f39870b4a855\") " pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.057951 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a1724655-9ac1-43dc-8292-f39870b4a855-webhook-cert\") pod \"metallb-operator-controller-manager-548f76d7c-s5g2f\" (UID: \"a1724655-9ac1-43dc-8292-f39870b4a855\") " pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.058160 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a1724655-9ac1-43dc-8292-f39870b4a855-apiservice-cert\") pod \"metallb-operator-controller-manager-548f76d7c-s5g2f\" (UID: \"a1724655-9ac1-43dc-8292-f39870b4a855\") " pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.159266 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a1724655-9ac1-43dc-8292-f39870b4a855-webhook-cert\") pod \"metallb-operator-controller-manager-548f76d7c-s5g2f\" (UID: \"a1724655-9ac1-43dc-8292-f39870b4a855\") " pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.159391 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a1724655-9ac1-43dc-8292-f39870b4a855-apiservice-cert\") pod \"metallb-operator-controller-manager-548f76d7c-s5g2f\" (UID: \"a1724655-9ac1-43dc-8292-f39870b4a855\") " pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.160417 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4j89\" (UniqueName: \"kubernetes.io/projected/a1724655-9ac1-43dc-8292-f39870b4a855-kube-api-access-g4j89\") pod \"metallb-operator-controller-manager-548f76d7c-s5g2f\" (UID: \"a1724655-9ac1-43dc-8292-f39870b4a855\") " pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.167383 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a1724655-9ac1-43dc-8292-f39870b4a855-webhook-cert\") pod \"metallb-operator-controller-manager-548f76d7c-s5g2f\" (UID: \"a1724655-9ac1-43dc-8292-f39870b4a855\") " pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.172002 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a1724655-9ac1-43dc-8292-f39870b4a855-apiservice-cert\") pod \"metallb-operator-controller-manager-548f76d7c-s5g2f\" (UID: \"a1724655-9ac1-43dc-8292-f39870b4a855\") " pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.187929 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4j89\" (UniqueName: \"kubernetes.io/projected/a1724655-9ac1-43dc-8292-f39870b4a855-kube-api-access-g4j89\") pod \"metallb-operator-controller-manager-548f76d7c-s5g2f\" (UID: \"a1724655-9ac1-43dc-8292-f39870b4a855\") " pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.267723 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd"] Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.268842 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.273896 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.274111 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.274357 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-xss8s" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.309541 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.311111 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd"] Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.366237 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcbtm\" (UniqueName: \"kubernetes.io/projected/77ecf1f1-1f2d-49a6-af38-32d16810ae9d-kube-api-access-fcbtm\") pod \"metallb-operator-webhook-server-867dbdb569-pqqsd\" (UID: \"77ecf1f1-1f2d-49a6-af38-32d16810ae9d\") " pod="metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.366339 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/77ecf1f1-1f2d-49a6-af38-32d16810ae9d-apiservice-cert\") pod \"metallb-operator-webhook-server-867dbdb569-pqqsd\" (UID: \"77ecf1f1-1f2d-49a6-af38-32d16810ae9d\") " pod="metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.366384 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/77ecf1f1-1f2d-49a6-af38-32d16810ae9d-webhook-cert\") pod \"metallb-operator-webhook-server-867dbdb569-pqqsd\" (UID: \"77ecf1f1-1f2d-49a6-af38-32d16810ae9d\") " pod="metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.467781 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcbtm\" (UniqueName: \"kubernetes.io/projected/77ecf1f1-1f2d-49a6-af38-32d16810ae9d-kube-api-access-fcbtm\") pod \"metallb-operator-webhook-server-867dbdb569-pqqsd\" (UID: \"77ecf1f1-1f2d-49a6-af38-32d16810ae9d\") " pod="metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.467890 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/77ecf1f1-1f2d-49a6-af38-32d16810ae9d-apiservice-cert\") pod \"metallb-operator-webhook-server-867dbdb569-pqqsd\" (UID: \"77ecf1f1-1f2d-49a6-af38-32d16810ae9d\") " pod="metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.467934 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/77ecf1f1-1f2d-49a6-af38-32d16810ae9d-webhook-cert\") pod \"metallb-operator-webhook-server-867dbdb569-pqqsd\" (UID: \"77ecf1f1-1f2d-49a6-af38-32d16810ae9d\") " pod="metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.476389 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/77ecf1f1-1f2d-49a6-af38-32d16810ae9d-apiservice-cert\") pod \"metallb-operator-webhook-server-867dbdb569-pqqsd\" (UID: \"77ecf1f1-1f2d-49a6-af38-32d16810ae9d\") " pod="metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.478061 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/77ecf1f1-1f2d-49a6-af38-32d16810ae9d-webhook-cert\") pod \"metallb-operator-webhook-server-867dbdb569-pqqsd\" (UID: \"77ecf1f1-1f2d-49a6-af38-32d16810ae9d\") " pod="metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.490945 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcbtm\" (UniqueName: \"kubernetes.io/projected/77ecf1f1-1f2d-49a6-af38-32d16810ae9d-kube-api-access-fcbtm\") pod \"metallb-operator-webhook-server-867dbdb569-pqqsd\" (UID: \"77ecf1f1-1f2d-49a6-af38-32d16810ae9d\") " pod="metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.586053 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f"] Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.596457 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd" Nov 25 16:59:16 crc kubenswrapper[4802]: I1125 16:59:16.800737 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd"] Nov 25 16:59:16 crc kubenswrapper[4802]: W1125 16:59:16.809202 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77ecf1f1_1f2d_49a6_af38_32d16810ae9d.slice/crio-157009c6dcb17d14ae011807fc4607501726d5eaf43348b6286f150633815f6a WatchSource:0}: Error finding container 157009c6dcb17d14ae011807fc4607501726d5eaf43348b6286f150633815f6a: Status 404 returned error can't find the container with id 157009c6dcb17d14ae011807fc4607501726d5eaf43348b6286f150633815f6a Nov 25 16:59:17 crc kubenswrapper[4802]: I1125 16:59:17.252987 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" event={"ID":"a1724655-9ac1-43dc-8292-f39870b4a855","Type":"ContainerStarted","Data":"d335d3aa33789cc3317ce0038a01d3bca67c9ee1a9588e6b98a4664125f519bf"} Nov 25 16:59:17 crc kubenswrapper[4802]: I1125 16:59:17.255276 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd" event={"ID":"77ecf1f1-1f2d-49a6-af38-32d16810ae9d","Type":"ContainerStarted","Data":"157009c6dcb17d14ae011807fc4607501726d5eaf43348b6286f150633815f6a"} Nov 25 16:59:24 crc kubenswrapper[4802]: I1125 16:59:24.248679 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:59:24 crc kubenswrapper[4802]: I1125 16:59:24.249567 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:59:24 crc kubenswrapper[4802]: I1125 16:59:24.306334 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" event={"ID":"a1724655-9ac1-43dc-8292-f39870b4a855","Type":"ContainerStarted","Data":"9d01284fca842bdc1379cf7add88c1dbce08a5591c1f4bb10ac4316ee241d9c5"} Nov 25 16:59:24 crc kubenswrapper[4802]: I1125 16:59:24.306496 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" Nov 25 16:59:24 crc kubenswrapper[4802]: I1125 16:59:24.308355 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd" event={"ID":"77ecf1f1-1f2d-49a6-af38-32d16810ae9d","Type":"ContainerStarted","Data":"afd49c883c8ac56c91dfe455a4710dfd20d062b9650e295a42dbba7a451b10a2"} Nov 25 16:59:24 crc kubenswrapper[4802]: I1125 16:59:24.308802 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd" Nov 25 16:59:24 crc kubenswrapper[4802]: I1125 16:59:24.331056 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" podStartSLOduration=2.332347294 podStartE2EDuration="9.331006313s" podCreationTimestamp="2025-11-25 16:59:15 +0000 UTC" firstStartedPulling="2025-11-25 16:59:16.598948573 +0000 UTC m=+739.743295759" lastFinishedPulling="2025-11-25 16:59:23.597607592 +0000 UTC m=+746.741954778" observedRunningTime="2025-11-25 16:59:24.329653237 +0000 UTC m=+747.474000423" watchObservedRunningTime="2025-11-25 16:59:24.331006313 +0000 UTC m=+747.475353499" Nov 25 16:59:24 crc kubenswrapper[4802]: I1125 16:59:24.349770 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd" podStartSLOduration=1.5459796080000001 podStartE2EDuration="8.34973514s" podCreationTimestamp="2025-11-25 16:59:16 +0000 UTC" firstStartedPulling="2025-11-25 16:59:16.812625309 +0000 UTC m=+739.956972495" lastFinishedPulling="2025-11-25 16:59:23.616380841 +0000 UTC m=+746.760728027" observedRunningTime="2025-11-25 16:59:24.348049525 +0000 UTC m=+747.492396721" watchObservedRunningTime="2025-11-25 16:59:24.34973514 +0000 UTC m=+747.494082336" Nov 25 16:59:25 crc kubenswrapper[4802]: I1125 16:59:25.739409 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bv6b6"] Nov 25 16:59:25 crc kubenswrapper[4802]: I1125 16:59:25.739820 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" podUID="49937863-524b-4ee0-b3c4-6522995947c4" containerName="controller-manager" containerID="cri-o://762effa9f3301816aee979f8a5a5131b381fbc426b98fa762ae9165bf218ba29" gracePeriod=30 Nov 25 16:59:25 crc kubenswrapper[4802]: I1125 16:59:25.826290 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w"] Nov 25 16:59:25 crc kubenswrapper[4802]: I1125 16:59:25.826903 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" podUID="ef963c3e-4e73-4fe3-aed1-21d8b7d8be40" containerName="route-controller-manager" containerID="cri-o://9a9233ef4628da8f5364a8d2d6c7ed3ab76ff73e1831a51c12d5a76019822c31" gracePeriod=30 Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.112907 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.174451 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.222908 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/49937863-524b-4ee0-b3c4-6522995947c4-proxy-ca-bundles\") pod \"49937863-524b-4ee0-b3c4-6522995947c4\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.222952 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-config\") pod \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\" (UID: \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\") " Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.223025 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4x99\" (UniqueName: \"kubernetes.io/projected/49937863-524b-4ee0-b3c4-6522995947c4-kube-api-access-t4x99\") pod \"49937863-524b-4ee0-b3c4-6522995947c4\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.223053 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49937863-524b-4ee0-b3c4-6522995947c4-config\") pod \"49937863-524b-4ee0-b3c4-6522995947c4\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.223097 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49937863-524b-4ee0-b3c4-6522995947c4-serving-cert\") pod \"49937863-524b-4ee0-b3c4-6522995947c4\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.223145 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4n8jz\" (UniqueName: \"kubernetes.io/projected/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-kube-api-access-4n8jz\") pod \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\" (UID: \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\") " Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.223187 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/49937863-524b-4ee0-b3c4-6522995947c4-client-ca\") pod \"49937863-524b-4ee0-b3c4-6522995947c4\" (UID: \"49937863-524b-4ee0-b3c4-6522995947c4\") " Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.223217 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-serving-cert\") pod \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\" (UID: \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\") " Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.223255 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-client-ca\") pod \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\" (UID: \"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40\") " Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.224165 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49937863-524b-4ee0-b3c4-6522995947c4-client-ca" (OuterVolumeSpecName: "client-ca") pod "49937863-524b-4ee0-b3c4-6522995947c4" (UID: "49937863-524b-4ee0-b3c4-6522995947c4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.224250 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-client-ca" (OuterVolumeSpecName: "client-ca") pod "ef963c3e-4e73-4fe3-aed1-21d8b7d8be40" (UID: "ef963c3e-4e73-4fe3-aed1-21d8b7d8be40"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.224405 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49937863-524b-4ee0-b3c4-6522995947c4-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "49937863-524b-4ee0-b3c4-6522995947c4" (UID: "49937863-524b-4ee0-b3c4-6522995947c4"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.224410 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-config" (OuterVolumeSpecName: "config") pod "ef963c3e-4e73-4fe3-aed1-21d8b7d8be40" (UID: "ef963c3e-4e73-4fe3-aed1-21d8b7d8be40"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.224434 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49937863-524b-4ee0-b3c4-6522995947c4-config" (OuterVolumeSpecName: "config") pod "49937863-524b-4ee0-b3c4-6522995947c4" (UID: "49937863-524b-4ee0-b3c4-6522995947c4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.231514 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49937863-524b-4ee0-b3c4-6522995947c4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "49937863-524b-4ee0-b3c4-6522995947c4" (UID: "49937863-524b-4ee0-b3c4-6522995947c4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.231516 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ef963c3e-4e73-4fe3-aed1-21d8b7d8be40" (UID: "ef963c3e-4e73-4fe3-aed1-21d8b7d8be40"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.231661 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49937863-524b-4ee0-b3c4-6522995947c4-kube-api-access-t4x99" (OuterVolumeSpecName: "kube-api-access-t4x99") pod "49937863-524b-4ee0-b3c4-6522995947c4" (UID: "49937863-524b-4ee0-b3c4-6522995947c4"). InnerVolumeSpecName "kube-api-access-t4x99". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.232807 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-kube-api-access-4n8jz" (OuterVolumeSpecName: "kube-api-access-4n8jz") pod "ef963c3e-4e73-4fe3-aed1-21d8b7d8be40" (UID: "ef963c3e-4e73-4fe3-aed1-21d8b7d8be40"). InnerVolumeSpecName "kube-api-access-4n8jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.327976 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4x99\" (UniqueName: \"kubernetes.io/projected/49937863-524b-4ee0-b3c4-6522995947c4-kube-api-access-t4x99\") on node \"crc\" DevicePath \"\"" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.328018 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49937863-524b-4ee0-b3c4-6522995947c4-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.328028 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49937863-524b-4ee0-b3c4-6522995947c4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.328041 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4n8jz\" (UniqueName: \"kubernetes.io/projected/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-kube-api-access-4n8jz\") on node \"crc\" DevicePath \"\"" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.328049 4802 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/49937863-524b-4ee0-b3c4-6522995947c4-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.328058 4802 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.328068 4802 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.328077 4802 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/49937863-524b-4ee0-b3c4-6522995947c4-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.328089 4802 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.331396 4802 generic.go:334] "Generic (PLEG): container finished" podID="49937863-524b-4ee0-b3c4-6522995947c4" containerID="762effa9f3301816aee979f8a5a5131b381fbc426b98fa762ae9165bf218ba29" exitCode=0 Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.331489 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" event={"ID":"49937863-524b-4ee0-b3c4-6522995947c4","Type":"ContainerDied","Data":"762effa9f3301816aee979f8a5a5131b381fbc426b98fa762ae9165bf218ba29"} Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.331601 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" event={"ID":"49937863-524b-4ee0-b3c4-6522995947c4","Type":"ContainerDied","Data":"bfdb812bd911c9d957a5ccbd08f626783b75cb2136935ae1c398341f5a3bcb8c"} Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.331626 4802 scope.go:117] "RemoveContainer" containerID="762effa9f3301816aee979f8a5a5131b381fbc426b98fa762ae9165bf218ba29" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.331909 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-bv6b6" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.337609 4802 generic.go:334] "Generic (PLEG): container finished" podID="ef963c3e-4e73-4fe3-aed1-21d8b7d8be40" containerID="9a9233ef4628da8f5364a8d2d6c7ed3ab76ff73e1831a51c12d5a76019822c31" exitCode=0 Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.338642 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.341385 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" event={"ID":"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40","Type":"ContainerDied","Data":"9a9233ef4628da8f5364a8d2d6c7ed3ab76ff73e1831a51c12d5a76019822c31"} Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.341467 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w" event={"ID":"ef963c3e-4e73-4fe3-aed1-21d8b7d8be40","Type":"ContainerDied","Data":"209d464547d064d8f2f3c862d07b0554d0f34b6a162ba999435821ef00f67dc3"} Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.358450 4802 scope.go:117] "RemoveContainer" containerID="762effa9f3301816aee979f8a5a5131b381fbc426b98fa762ae9165bf218ba29" Nov 25 16:59:26 crc kubenswrapper[4802]: E1125 16:59:26.363193 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"762effa9f3301816aee979f8a5a5131b381fbc426b98fa762ae9165bf218ba29\": container with ID starting with 762effa9f3301816aee979f8a5a5131b381fbc426b98fa762ae9165bf218ba29 not found: ID does not exist" containerID="762effa9f3301816aee979f8a5a5131b381fbc426b98fa762ae9165bf218ba29" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.363258 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"762effa9f3301816aee979f8a5a5131b381fbc426b98fa762ae9165bf218ba29"} err="failed to get container status \"762effa9f3301816aee979f8a5a5131b381fbc426b98fa762ae9165bf218ba29\": rpc error: code = NotFound desc = could not find container \"762effa9f3301816aee979f8a5a5131b381fbc426b98fa762ae9165bf218ba29\": container with ID starting with 762effa9f3301816aee979f8a5a5131b381fbc426b98fa762ae9165bf218ba29 not found: ID does not exist" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.363290 4802 scope.go:117] "RemoveContainer" containerID="9a9233ef4628da8f5364a8d2d6c7ed3ab76ff73e1831a51c12d5a76019822c31" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.467633 4802 scope.go:117] "RemoveContainer" containerID="9a9233ef4628da8f5364a8d2d6c7ed3ab76ff73e1831a51c12d5a76019822c31" Nov 25 16:59:26 crc kubenswrapper[4802]: E1125 16:59:26.468228 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a9233ef4628da8f5364a8d2d6c7ed3ab76ff73e1831a51c12d5a76019822c31\": container with ID starting with 9a9233ef4628da8f5364a8d2d6c7ed3ab76ff73e1831a51c12d5a76019822c31 not found: ID does not exist" containerID="9a9233ef4628da8f5364a8d2d6c7ed3ab76ff73e1831a51c12d5a76019822c31" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.468263 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a9233ef4628da8f5364a8d2d6c7ed3ab76ff73e1831a51c12d5a76019822c31"} err="failed to get container status \"9a9233ef4628da8f5364a8d2d6c7ed3ab76ff73e1831a51c12d5a76019822c31\": rpc error: code = NotFound desc = could not find container \"9a9233ef4628da8f5364a8d2d6c7ed3ab76ff73e1831a51c12d5a76019822c31\": container with ID starting with 9a9233ef4628da8f5364a8d2d6c7ed3ab76ff73e1831a51c12d5a76019822c31 not found: ID does not exist" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.482080 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w"] Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.492708 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zvb9w"] Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.510056 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bv6b6"] Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.515509 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bv6b6"] Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.529726 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-76997f8cd8-h46c4"] Nov 25 16:59:26 crc kubenswrapper[4802]: E1125 16:59:26.529989 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef963c3e-4e73-4fe3-aed1-21d8b7d8be40" containerName="route-controller-manager" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.530005 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef963c3e-4e73-4fe3-aed1-21d8b7d8be40" containerName="route-controller-manager" Nov 25 16:59:26 crc kubenswrapper[4802]: E1125 16:59:26.530019 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49937863-524b-4ee0-b3c4-6522995947c4" containerName="controller-manager" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.530024 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="49937863-524b-4ee0-b3c4-6522995947c4" containerName="controller-manager" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.530144 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef963c3e-4e73-4fe3-aed1-21d8b7d8be40" containerName="route-controller-manager" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.530156 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="49937863-524b-4ee0-b3c4-6522995947c4" containerName="controller-manager" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.530586 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.543087 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.543637 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.545932 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.546198 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.546414 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.546549 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.560187 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 16:59:26 crc kubenswrapper[4802]: E1125 16:59:26.568693 4802 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef963c3e_4e73_4fe3_aed1_21d8b7d8be40.slice/crio-209d464547d064d8f2f3c862d07b0554d0f34b6a162ba999435821ef00f67dc3\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef963c3e_4e73_4fe3_aed1_21d8b7d8be40.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49937863_524b_4ee0_b3c4_6522995947c4.slice/crio-bfdb812bd911c9d957a5ccbd08f626783b75cb2136935ae1c398341f5a3bcb8c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49937863_524b_4ee0_b3c4_6522995947c4.slice\": RecentStats: unable to find data in memory cache]" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.575535 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-76997f8cd8-h46c4"] Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.602205 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7"] Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.603052 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.616654 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7"] Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.617374 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.617468 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.617630 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.618400 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.618545 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.618679 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.634021 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/44c902cc-e2a0-4653-9435-5cd5f2ccd631-client-ca\") pod \"controller-manager-76997f8cd8-h46c4\" (UID: \"44c902cc-e2a0-4653-9435-5cd5f2ccd631\") " pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.634099 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d92a19a5-6b76-45a2-a079-a458cbc285ca-serving-cert\") pod \"route-controller-manager-596579d78c-mljc7\" (UID: \"d92a19a5-6b76-45a2-a079-a458cbc285ca\") " pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.634153 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk59p\" (UniqueName: \"kubernetes.io/projected/d92a19a5-6b76-45a2-a079-a458cbc285ca-kube-api-access-wk59p\") pod \"route-controller-manager-596579d78c-mljc7\" (UID: \"d92a19a5-6b76-45a2-a079-a458cbc285ca\") " pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.634178 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/44c902cc-e2a0-4653-9435-5cd5f2ccd631-proxy-ca-bundles\") pod \"controller-manager-76997f8cd8-h46c4\" (UID: \"44c902cc-e2a0-4653-9435-5cd5f2ccd631\") " pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.634193 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d92a19a5-6b76-45a2-a079-a458cbc285ca-client-ca\") pod \"route-controller-manager-596579d78c-mljc7\" (UID: \"d92a19a5-6b76-45a2-a079-a458cbc285ca\") " pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.634211 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44c902cc-e2a0-4653-9435-5cd5f2ccd631-config\") pod \"controller-manager-76997f8cd8-h46c4\" (UID: \"44c902cc-e2a0-4653-9435-5cd5f2ccd631\") " pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.634251 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44c902cc-e2a0-4653-9435-5cd5f2ccd631-serving-cert\") pod \"controller-manager-76997f8cd8-h46c4\" (UID: \"44c902cc-e2a0-4653-9435-5cd5f2ccd631\") " pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.634484 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d92a19a5-6b76-45a2-a079-a458cbc285ca-config\") pod \"route-controller-manager-596579d78c-mljc7\" (UID: \"d92a19a5-6b76-45a2-a079-a458cbc285ca\") " pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.634507 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96blk\" (UniqueName: \"kubernetes.io/projected/44c902cc-e2a0-4653-9435-5cd5f2ccd631-kube-api-access-96blk\") pod \"controller-manager-76997f8cd8-h46c4\" (UID: \"44c902cc-e2a0-4653-9435-5cd5f2ccd631\") " pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.735735 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96blk\" (UniqueName: \"kubernetes.io/projected/44c902cc-e2a0-4653-9435-5cd5f2ccd631-kube-api-access-96blk\") pod \"controller-manager-76997f8cd8-h46c4\" (UID: \"44c902cc-e2a0-4653-9435-5cd5f2ccd631\") " pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.735826 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/44c902cc-e2a0-4653-9435-5cd5f2ccd631-client-ca\") pod \"controller-manager-76997f8cd8-h46c4\" (UID: \"44c902cc-e2a0-4653-9435-5cd5f2ccd631\") " pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.735849 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d92a19a5-6b76-45a2-a079-a458cbc285ca-serving-cert\") pod \"route-controller-manager-596579d78c-mljc7\" (UID: \"d92a19a5-6b76-45a2-a079-a458cbc285ca\") " pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.735867 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk59p\" (UniqueName: \"kubernetes.io/projected/d92a19a5-6b76-45a2-a079-a458cbc285ca-kube-api-access-wk59p\") pod \"route-controller-manager-596579d78c-mljc7\" (UID: \"d92a19a5-6b76-45a2-a079-a458cbc285ca\") " pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.735890 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d92a19a5-6b76-45a2-a079-a458cbc285ca-client-ca\") pod \"route-controller-manager-596579d78c-mljc7\" (UID: \"d92a19a5-6b76-45a2-a079-a458cbc285ca\") " pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.735904 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/44c902cc-e2a0-4653-9435-5cd5f2ccd631-proxy-ca-bundles\") pod \"controller-manager-76997f8cd8-h46c4\" (UID: \"44c902cc-e2a0-4653-9435-5cd5f2ccd631\") " pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.735922 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44c902cc-e2a0-4653-9435-5cd5f2ccd631-config\") pod \"controller-manager-76997f8cd8-h46c4\" (UID: \"44c902cc-e2a0-4653-9435-5cd5f2ccd631\") " pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.735957 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44c902cc-e2a0-4653-9435-5cd5f2ccd631-serving-cert\") pod \"controller-manager-76997f8cd8-h46c4\" (UID: \"44c902cc-e2a0-4653-9435-5cd5f2ccd631\") " pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.735992 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d92a19a5-6b76-45a2-a079-a458cbc285ca-config\") pod \"route-controller-manager-596579d78c-mljc7\" (UID: \"d92a19a5-6b76-45a2-a079-a458cbc285ca\") " pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.737789 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d92a19a5-6b76-45a2-a079-a458cbc285ca-config\") pod \"route-controller-manager-596579d78c-mljc7\" (UID: \"d92a19a5-6b76-45a2-a079-a458cbc285ca\") " pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.738004 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/44c902cc-e2a0-4653-9435-5cd5f2ccd631-proxy-ca-bundles\") pod \"controller-manager-76997f8cd8-h46c4\" (UID: \"44c902cc-e2a0-4653-9435-5cd5f2ccd631\") " pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.738266 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/44c902cc-e2a0-4653-9435-5cd5f2ccd631-client-ca\") pod \"controller-manager-76997f8cd8-h46c4\" (UID: \"44c902cc-e2a0-4653-9435-5cd5f2ccd631\") " pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.738300 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d92a19a5-6b76-45a2-a079-a458cbc285ca-client-ca\") pod \"route-controller-manager-596579d78c-mljc7\" (UID: \"d92a19a5-6b76-45a2-a079-a458cbc285ca\") " pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.738353 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44c902cc-e2a0-4653-9435-5cd5f2ccd631-config\") pod \"controller-manager-76997f8cd8-h46c4\" (UID: \"44c902cc-e2a0-4653-9435-5cd5f2ccd631\") " pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.741146 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44c902cc-e2a0-4653-9435-5cd5f2ccd631-serving-cert\") pod \"controller-manager-76997f8cd8-h46c4\" (UID: \"44c902cc-e2a0-4653-9435-5cd5f2ccd631\") " pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.741146 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d92a19a5-6b76-45a2-a079-a458cbc285ca-serving-cert\") pod \"route-controller-manager-596579d78c-mljc7\" (UID: \"d92a19a5-6b76-45a2-a079-a458cbc285ca\") " pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.758456 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96blk\" (UniqueName: \"kubernetes.io/projected/44c902cc-e2a0-4653-9435-5cd5f2ccd631-kube-api-access-96blk\") pod \"controller-manager-76997f8cd8-h46c4\" (UID: \"44c902cc-e2a0-4653-9435-5cd5f2ccd631\") " pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.758916 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk59p\" (UniqueName: \"kubernetes.io/projected/d92a19a5-6b76-45a2-a079-a458cbc285ca-kube-api-access-wk59p\") pod \"route-controller-manager-596579d78c-mljc7\" (UID: \"d92a19a5-6b76-45a2-a079-a458cbc285ca\") " pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.855899 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:26 crc kubenswrapper[4802]: I1125 16:59:26.933974 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" Nov 25 16:59:27 crc kubenswrapper[4802]: I1125 16:59:27.096538 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-76997f8cd8-h46c4"] Nov 25 16:59:27 crc kubenswrapper[4802]: I1125 16:59:27.171227 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7"] Nov 25 16:59:27 crc kubenswrapper[4802]: W1125 16:59:27.183117 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd92a19a5_6b76_45a2_a079_a458cbc285ca.slice/crio-8fec42bacc516676c792b1823d4aaa2631ba16379229338c9eaaa75ec1bfcb68 WatchSource:0}: Error finding container 8fec42bacc516676c792b1823d4aaa2631ba16379229338c9eaaa75ec1bfcb68: Status 404 returned error can't find the container with id 8fec42bacc516676c792b1823d4aaa2631ba16379229338c9eaaa75ec1bfcb68 Nov 25 16:59:27 crc kubenswrapper[4802]: I1125 16:59:27.344046 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" event={"ID":"d92a19a5-6b76-45a2-a079-a458cbc285ca","Type":"ContainerStarted","Data":"8fec42bacc516676c792b1823d4aaa2631ba16379229338c9eaaa75ec1bfcb68"} Nov 25 16:59:27 crc kubenswrapper[4802]: I1125 16:59:27.345409 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" event={"ID":"44c902cc-e2a0-4653-9435-5cd5f2ccd631","Type":"ContainerStarted","Data":"d4cf5fa21687fe5f3923017bc51230d392052e5e01f92f3bde749e4afa62f3a5"} Nov 25 16:59:27 crc kubenswrapper[4802]: I1125 16:59:27.511605 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49937863-524b-4ee0-b3c4-6522995947c4" path="/var/lib/kubelet/pods/49937863-524b-4ee0-b3c4-6522995947c4/volumes" Nov 25 16:59:27 crc kubenswrapper[4802]: I1125 16:59:27.512436 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef963c3e-4e73-4fe3-aed1-21d8b7d8be40" path="/var/lib/kubelet/pods/ef963c3e-4e73-4fe3-aed1-21d8b7d8be40/volumes" Nov 25 16:59:28 crc kubenswrapper[4802]: I1125 16:59:28.354390 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" event={"ID":"44c902cc-e2a0-4653-9435-5cd5f2ccd631","Type":"ContainerStarted","Data":"4c827db9cd15d0729b7a9c536934f9095f5b76ac3f460f895d2bc5f353b8e38f"} Nov 25 16:59:29 crc kubenswrapper[4802]: I1125 16:59:29.361944 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" event={"ID":"d92a19a5-6b76-45a2-a079-a458cbc285ca","Type":"ContainerStarted","Data":"0ccd824000d78614ff35ebe398fc2c467b2a6cb9997a378761d7df8f13e66d0f"} Nov 25 16:59:29 crc kubenswrapper[4802]: I1125 16:59:29.362473 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:29 crc kubenswrapper[4802]: I1125 16:59:29.362519 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" Nov 25 16:59:29 crc kubenswrapper[4802]: I1125 16:59:29.366648 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" Nov 25 16:59:29 crc kubenswrapper[4802]: I1125 16:59:29.382299 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-76997f8cd8-h46c4" podStartSLOduration=3.3822774239999998 podStartE2EDuration="3.382277424s" podCreationTimestamp="2025-11-25 16:59:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:59:29.381535814 +0000 UTC m=+752.525883010" watchObservedRunningTime="2025-11-25 16:59:29.382277424 +0000 UTC m=+752.526624610" Nov 25 16:59:29 crc kubenswrapper[4802]: I1125 16:59:29.453482 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" podStartSLOduration=3.453460562 podStartE2EDuration="3.453460562s" podCreationTimestamp="2025-11-25 16:59:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:59:29.452476405 +0000 UTC m=+752.596823611" watchObservedRunningTime="2025-11-25 16:59:29.453460562 +0000 UTC m=+752.597807748" Nov 25 16:59:29 crc kubenswrapper[4802]: I1125 16:59:29.635333 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-596579d78c-mljc7" Nov 25 16:59:32 crc kubenswrapper[4802]: I1125 16:59:32.497985 4802 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 16:59:36 crc kubenswrapper[4802]: I1125 16:59:36.627293 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-867dbdb569-pqqsd" Nov 25 16:59:54 crc kubenswrapper[4802]: I1125 16:59:54.248685 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:59:54 crc kubenswrapper[4802]: I1125 16:59:54.249604 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:59:54 crc kubenswrapper[4802]: I1125 16:59:54.249679 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 16:59:54 crc kubenswrapper[4802]: I1125 16:59:54.250403 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2dd6a8bbe96b494880e307bbcf47128d8479e3413bfeccced18011d7c6e2daae"} pod="openshift-machine-config-operator/machine-config-daemon-h29wc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 16:59:54 crc kubenswrapper[4802]: I1125 16:59:54.250588 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" containerID="cri-o://2dd6a8bbe96b494880e307bbcf47128d8479e3413bfeccced18011d7c6e2daae" gracePeriod=600 Nov 25 16:59:55 crc kubenswrapper[4802]: I1125 16:59:55.524042 4802 generic.go:334] "Generic (PLEG): container finished" podID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerID="2dd6a8bbe96b494880e307bbcf47128d8479e3413bfeccced18011d7c6e2daae" exitCode=0 Nov 25 16:59:55 crc kubenswrapper[4802]: I1125 16:59:55.524133 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerDied","Data":"2dd6a8bbe96b494880e307bbcf47128d8479e3413bfeccced18011d7c6e2daae"} Nov 25 16:59:55 crc kubenswrapper[4802]: I1125 16:59:55.524664 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerStarted","Data":"b903688e55473d1e5bdabb3f4b23a6b1ac78c87eaabad518c885de28866fe47e"} Nov 25 16:59:55 crc kubenswrapper[4802]: I1125 16:59:55.524710 4802 scope.go:117] "RemoveContainer" containerID="3d37aca18ca92000cdaea6d946fa2206f34b1d2780852086c38dfe3898b186c8" Nov 25 16:59:56 crc kubenswrapper[4802]: I1125 16:59:56.314139 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.082167 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-rzq2t"] Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.085542 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.090550 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-rzknk"] Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.090974 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-8dk78" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.091234 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.091263 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.097828 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-rzknk" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.104436 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-rzknk"] Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.107409 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.168603 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aee28310-d76a-491b-b174-e1d131d078c7-metrics-certs\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.168661 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d605e3c4-111d-4f35-bf63-f855001d1967-cert\") pod \"frr-k8s-webhook-server-6998585d5-rzknk\" (UID: \"d605e3c4-111d-4f35-bf63-f855001d1967\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-rzknk" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.168697 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/aee28310-d76a-491b-b174-e1d131d078c7-reloader\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.168742 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/aee28310-d76a-491b-b174-e1d131d078c7-frr-conf\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.168774 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/aee28310-d76a-491b-b174-e1d131d078c7-frr-sockets\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.168798 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/aee28310-d76a-491b-b174-e1d131d078c7-frr-startup\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.168853 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dnbd\" (UniqueName: \"kubernetes.io/projected/aee28310-d76a-491b-b174-e1d131d078c7-kube-api-access-7dnbd\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.168887 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/aee28310-d76a-491b-b174-e1d131d078c7-metrics\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.168905 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4bcs\" (UniqueName: \"kubernetes.io/projected/d605e3c4-111d-4f35-bf63-f855001d1967-kube-api-access-m4bcs\") pod \"frr-k8s-webhook-server-6998585d5-rzknk\" (UID: \"d605e3c4-111d-4f35-bf63-f855001d1967\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-rzknk" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.189127 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-xnshj"] Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.190486 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-xnshj" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.193680 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.194144 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.194906 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-8z52s" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.198460 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.200433 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-tcf29"] Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.201773 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-tcf29" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.203510 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.218637 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-tcf29"] Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.269924 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/aee28310-d76a-491b-b174-e1d131d078c7-frr-conf\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.269990 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk69n\" (UniqueName: \"kubernetes.io/projected/cd0861df-163a-47f7-8a35-8b8c25e02ca4-kube-api-access-nk69n\") pod \"speaker-xnshj\" (UID: \"cd0861df-163a-47f7-8a35-8b8c25e02ca4\") " pod="metallb-system/speaker-xnshj" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.270033 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/aee28310-d76a-491b-b174-e1d131d078c7-frr-sockets\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.270059 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/85895137-3ed8-4214-92fa-ea23980c5083-cert\") pod \"controller-6c7b4b5f48-tcf29\" (UID: \"85895137-3ed8-4214-92fa-ea23980c5083\") " pod="metallb-system/controller-6c7b4b5f48-tcf29" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.270086 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/aee28310-d76a-491b-b174-e1d131d078c7-frr-startup\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.270105 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dnbd\" (UniqueName: \"kubernetes.io/projected/aee28310-d76a-491b-b174-e1d131d078c7-kube-api-access-7dnbd\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.270145 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/aee28310-d76a-491b-b174-e1d131d078c7-metrics\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.270171 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjqhp\" (UniqueName: \"kubernetes.io/projected/85895137-3ed8-4214-92fa-ea23980c5083-kube-api-access-gjqhp\") pod \"controller-6c7b4b5f48-tcf29\" (UID: \"85895137-3ed8-4214-92fa-ea23980c5083\") " pod="metallb-system/controller-6c7b4b5f48-tcf29" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.270191 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4bcs\" (UniqueName: \"kubernetes.io/projected/d605e3c4-111d-4f35-bf63-f855001d1967-kube-api-access-m4bcs\") pod \"frr-k8s-webhook-server-6998585d5-rzknk\" (UID: \"d605e3c4-111d-4f35-bf63-f855001d1967\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-rzknk" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.270208 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/cd0861df-163a-47f7-8a35-8b8c25e02ca4-memberlist\") pod \"speaker-xnshj\" (UID: \"cd0861df-163a-47f7-8a35-8b8c25e02ca4\") " pod="metallb-system/speaker-xnshj" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.270229 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aee28310-d76a-491b-b174-e1d131d078c7-metrics-certs\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.270247 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d605e3c4-111d-4f35-bf63-f855001d1967-cert\") pod \"frr-k8s-webhook-server-6998585d5-rzknk\" (UID: \"d605e3c4-111d-4f35-bf63-f855001d1967\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-rzknk" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.270262 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cd0861df-163a-47f7-8a35-8b8c25e02ca4-metrics-certs\") pod \"speaker-xnshj\" (UID: \"cd0861df-163a-47f7-8a35-8b8c25e02ca4\") " pod="metallb-system/speaker-xnshj" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.270275 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/cd0861df-163a-47f7-8a35-8b8c25e02ca4-metallb-excludel2\") pod \"speaker-xnshj\" (UID: \"cd0861df-163a-47f7-8a35-8b8c25e02ca4\") " pod="metallb-system/speaker-xnshj" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.270300 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/aee28310-d76a-491b-b174-e1d131d078c7-reloader\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.270331 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/85895137-3ed8-4214-92fa-ea23980c5083-metrics-certs\") pod \"controller-6c7b4b5f48-tcf29\" (UID: \"85895137-3ed8-4214-92fa-ea23980c5083\") " pod="metallb-system/controller-6c7b4b5f48-tcf29" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.270517 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/aee28310-d76a-491b-b174-e1d131d078c7-frr-conf\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.270752 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/aee28310-d76a-491b-b174-e1d131d078c7-metrics\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.271068 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/aee28310-d76a-491b-b174-e1d131d078c7-reloader\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.271162 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/aee28310-d76a-491b-b174-e1d131d078c7-frr-sockets\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.272122 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/aee28310-d76a-491b-b174-e1d131d078c7-frr-startup\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.278717 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aee28310-d76a-491b-b174-e1d131d078c7-metrics-certs\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.280308 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d605e3c4-111d-4f35-bf63-f855001d1967-cert\") pod \"frr-k8s-webhook-server-6998585d5-rzknk\" (UID: \"d605e3c4-111d-4f35-bf63-f855001d1967\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-rzknk" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.288299 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4bcs\" (UniqueName: \"kubernetes.io/projected/d605e3c4-111d-4f35-bf63-f855001d1967-kube-api-access-m4bcs\") pod \"frr-k8s-webhook-server-6998585d5-rzknk\" (UID: \"d605e3c4-111d-4f35-bf63-f855001d1967\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-rzknk" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.291564 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dnbd\" (UniqueName: \"kubernetes.io/projected/aee28310-d76a-491b-b174-e1d131d078c7-kube-api-access-7dnbd\") pod \"frr-k8s-rzq2t\" (UID: \"aee28310-d76a-491b-b174-e1d131d078c7\") " pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.371281 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/85895137-3ed8-4214-92fa-ea23980c5083-metrics-certs\") pod \"controller-6c7b4b5f48-tcf29\" (UID: \"85895137-3ed8-4214-92fa-ea23980c5083\") " pod="metallb-system/controller-6c7b4b5f48-tcf29" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.371366 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk69n\" (UniqueName: \"kubernetes.io/projected/cd0861df-163a-47f7-8a35-8b8c25e02ca4-kube-api-access-nk69n\") pod \"speaker-xnshj\" (UID: \"cd0861df-163a-47f7-8a35-8b8c25e02ca4\") " pod="metallb-system/speaker-xnshj" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.371387 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/85895137-3ed8-4214-92fa-ea23980c5083-cert\") pod \"controller-6c7b4b5f48-tcf29\" (UID: \"85895137-3ed8-4214-92fa-ea23980c5083\") " pod="metallb-system/controller-6c7b4b5f48-tcf29" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.371425 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjqhp\" (UniqueName: \"kubernetes.io/projected/85895137-3ed8-4214-92fa-ea23980c5083-kube-api-access-gjqhp\") pod \"controller-6c7b4b5f48-tcf29\" (UID: \"85895137-3ed8-4214-92fa-ea23980c5083\") " pod="metallb-system/controller-6c7b4b5f48-tcf29" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.371449 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/cd0861df-163a-47f7-8a35-8b8c25e02ca4-memberlist\") pod \"speaker-xnshj\" (UID: \"cd0861df-163a-47f7-8a35-8b8c25e02ca4\") " pod="metallb-system/speaker-xnshj" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.371473 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cd0861df-163a-47f7-8a35-8b8c25e02ca4-metrics-certs\") pod \"speaker-xnshj\" (UID: \"cd0861df-163a-47f7-8a35-8b8c25e02ca4\") " pod="metallb-system/speaker-xnshj" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.371489 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/cd0861df-163a-47f7-8a35-8b8c25e02ca4-metallb-excludel2\") pod \"speaker-xnshj\" (UID: \"cd0861df-163a-47f7-8a35-8b8c25e02ca4\") " pod="metallb-system/speaker-xnshj" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.372286 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/cd0861df-163a-47f7-8a35-8b8c25e02ca4-metallb-excludel2\") pod \"speaker-xnshj\" (UID: \"cd0861df-163a-47f7-8a35-8b8c25e02ca4\") " pod="metallb-system/speaker-xnshj" Nov 25 16:59:57 crc kubenswrapper[4802]: E1125 16:59:57.372437 4802 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 16:59:57 crc kubenswrapper[4802]: E1125 16:59:57.372541 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd0861df-163a-47f7-8a35-8b8c25e02ca4-memberlist podName:cd0861df-163a-47f7-8a35-8b8c25e02ca4 nodeName:}" failed. No retries permitted until 2025-11-25 16:59:57.872515883 +0000 UTC m=+781.016863149 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/cd0861df-163a-47f7-8a35-8b8c25e02ca4-memberlist") pod "speaker-xnshj" (UID: "cd0861df-163a-47f7-8a35-8b8c25e02ca4") : secret "metallb-memberlist" not found Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.375502 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cd0861df-163a-47f7-8a35-8b8c25e02ca4-metrics-certs\") pod \"speaker-xnshj\" (UID: \"cd0861df-163a-47f7-8a35-8b8c25e02ca4\") " pod="metallb-system/speaker-xnshj" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.375978 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/85895137-3ed8-4214-92fa-ea23980c5083-cert\") pod \"controller-6c7b4b5f48-tcf29\" (UID: \"85895137-3ed8-4214-92fa-ea23980c5083\") " pod="metallb-system/controller-6c7b4b5f48-tcf29" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.377041 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/85895137-3ed8-4214-92fa-ea23980c5083-metrics-certs\") pod \"controller-6c7b4b5f48-tcf29\" (UID: \"85895137-3ed8-4214-92fa-ea23980c5083\") " pod="metallb-system/controller-6c7b4b5f48-tcf29" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.389572 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk69n\" (UniqueName: \"kubernetes.io/projected/cd0861df-163a-47f7-8a35-8b8c25e02ca4-kube-api-access-nk69n\") pod \"speaker-xnshj\" (UID: \"cd0861df-163a-47f7-8a35-8b8c25e02ca4\") " pod="metallb-system/speaker-xnshj" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.390736 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjqhp\" (UniqueName: \"kubernetes.io/projected/85895137-3ed8-4214-92fa-ea23980c5083-kube-api-access-gjqhp\") pod \"controller-6c7b4b5f48-tcf29\" (UID: \"85895137-3ed8-4214-92fa-ea23980c5083\") " pod="metallb-system/controller-6c7b4b5f48-tcf29" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.415534 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-rzq2t" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.427505 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-rzknk" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.525779 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-tcf29" Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.854573 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-rzknk"] Nov 25 16:59:57 crc kubenswrapper[4802]: W1125 16:59:57.857773 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd605e3c4_111d_4f35_bf63_f855001d1967.slice/crio-bf870d493b8740a0e8c9e6d744390d1f000e51b7f067dbc41bc018c04479317b WatchSource:0}: Error finding container bf870d493b8740a0e8c9e6d744390d1f000e51b7f067dbc41bc018c04479317b: Status 404 returned error can't find the container with id bf870d493b8740a0e8c9e6d744390d1f000e51b7f067dbc41bc018c04479317b Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.880096 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/cd0861df-163a-47f7-8a35-8b8c25e02ca4-memberlist\") pod \"speaker-xnshj\" (UID: \"cd0861df-163a-47f7-8a35-8b8c25e02ca4\") " pod="metallb-system/speaker-xnshj" Nov 25 16:59:57 crc kubenswrapper[4802]: E1125 16:59:57.880370 4802 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 16:59:57 crc kubenswrapper[4802]: E1125 16:59:57.880437 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd0861df-163a-47f7-8a35-8b8c25e02ca4-memberlist podName:cd0861df-163a-47f7-8a35-8b8c25e02ca4 nodeName:}" failed. No retries permitted until 2025-11-25 16:59:58.880417234 +0000 UTC m=+782.024764420 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/cd0861df-163a-47f7-8a35-8b8c25e02ca4-memberlist") pod "speaker-xnshj" (UID: "cd0861df-163a-47f7-8a35-8b8c25e02ca4") : secret "metallb-memberlist" not found Nov 25 16:59:57 crc kubenswrapper[4802]: I1125 16:59:57.951446 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-tcf29"] Nov 25 16:59:57 crc kubenswrapper[4802]: W1125 16:59:57.956355 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85895137_3ed8_4214_92fa_ea23980c5083.slice/crio-699b07fe2825c7ea0c4f0f182d8451f04cb636b52aa7ee21e6e1b7c7022b0c4c WatchSource:0}: Error finding container 699b07fe2825c7ea0c4f0f182d8451f04cb636b52aa7ee21e6e1b7c7022b0c4c: Status 404 returned error can't find the container with id 699b07fe2825c7ea0c4f0f182d8451f04cb636b52aa7ee21e6e1b7c7022b0c4c Nov 25 16:59:58 crc kubenswrapper[4802]: I1125 16:59:58.558029 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rzq2t" event={"ID":"aee28310-d76a-491b-b174-e1d131d078c7","Type":"ContainerStarted","Data":"8a6d7d347488228e1469fb8d1f15e32ab37efb88131b77aea25c966a82cec722"} Nov 25 16:59:58 crc kubenswrapper[4802]: I1125 16:59:58.559656 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-rzknk" event={"ID":"d605e3c4-111d-4f35-bf63-f855001d1967","Type":"ContainerStarted","Data":"bf870d493b8740a0e8c9e6d744390d1f000e51b7f067dbc41bc018c04479317b"} Nov 25 16:59:58 crc kubenswrapper[4802]: I1125 16:59:58.562288 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-tcf29" event={"ID":"85895137-3ed8-4214-92fa-ea23980c5083","Type":"ContainerStarted","Data":"49f36de1556aba9002879bbc1e34f7477a390cfb97441c6e2678343891959bcb"} Nov 25 16:59:58 crc kubenswrapper[4802]: I1125 16:59:58.562316 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-tcf29" event={"ID":"85895137-3ed8-4214-92fa-ea23980c5083","Type":"ContainerStarted","Data":"699b07fe2825c7ea0c4f0f182d8451f04cb636b52aa7ee21e6e1b7c7022b0c4c"} Nov 25 16:59:58 crc kubenswrapper[4802]: I1125 16:59:58.895690 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/cd0861df-163a-47f7-8a35-8b8c25e02ca4-memberlist\") pod \"speaker-xnshj\" (UID: \"cd0861df-163a-47f7-8a35-8b8c25e02ca4\") " pod="metallb-system/speaker-xnshj" Nov 25 16:59:58 crc kubenswrapper[4802]: I1125 16:59:58.903861 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/cd0861df-163a-47f7-8a35-8b8c25e02ca4-memberlist\") pod \"speaker-xnshj\" (UID: \"cd0861df-163a-47f7-8a35-8b8c25e02ca4\") " pod="metallb-system/speaker-xnshj" Nov 25 16:59:59 crc kubenswrapper[4802]: I1125 16:59:59.009959 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-8z52s" Nov 25 16:59:59 crc kubenswrapper[4802]: I1125 16:59:59.018937 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-xnshj" Nov 25 16:59:59 crc kubenswrapper[4802]: W1125 16:59:59.039022 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd0861df_163a_47f7_8a35_8b8c25e02ca4.slice/crio-d671413376f741848c71a2e489fb05a4067e60f0ce831a912cac8056a1120407 WatchSource:0}: Error finding container d671413376f741848c71a2e489fb05a4067e60f0ce831a912cac8056a1120407: Status 404 returned error can't find the container with id d671413376f741848c71a2e489fb05a4067e60f0ce831a912cac8056a1120407 Nov 25 16:59:59 crc kubenswrapper[4802]: I1125 16:59:59.578601 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-xnshj" event={"ID":"cd0861df-163a-47f7-8a35-8b8c25e02ca4","Type":"ContainerStarted","Data":"ba1401dd1304e5b1b9900800a98108c630302c6e681776baf0aefed2f21f1a4a"} Nov 25 16:59:59 crc kubenswrapper[4802]: I1125 16:59:59.579106 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-xnshj" event={"ID":"cd0861df-163a-47f7-8a35-8b8c25e02ca4","Type":"ContainerStarted","Data":"d671413376f741848c71a2e489fb05a4067e60f0ce831a912cac8056a1120407"} Nov 25 17:00:00 crc kubenswrapper[4802]: I1125 17:00:00.157070 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks"] Nov 25 17:00:00 crc kubenswrapper[4802]: I1125 17:00:00.158545 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks" Nov 25 17:00:00 crc kubenswrapper[4802]: I1125 17:00:00.161632 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 17:00:00 crc kubenswrapper[4802]: I1125 17:00:00.161720 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 17:00:00 crc kubenswrapper[4802]: I1125 17:00:00.181192 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks"] Nov 25 17:00:00 crc kubenswrapper[4802]: I1125 17:00:00.323048 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z29mj\" (UniqueName: \"kubernetes.io/projected/2380f347-95ad-4e2f-b5c9-eda77578272b-kube-api-access-z29mj\") pod \"collect-profiles-29401500-4q2ks\" (UID: \"2380f347-95ad-4e2f-b5c9-eda77578272b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks" Nov 25 17:00:00 crc kubenswrapper[4802]: I1125 17:00:00.323109 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2380f347-95ad-4e2f-b5c9-eda77578272b-config-volume\") pod \"collect-profiles-29401500-4q2ks\" (UID: \"2380f347-95ad-4e2f-b5c9-eda77578272b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks" Nov 25 17:00:00 crc kubenswrapper[4802]: I1125 17:00:00.323170 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2380f347-95ad-4e2f-b5c9-eda77578272b-secret-volume\") pod \"collect-profiles-29401500-4q2ks\" (UID: \"2380f347-95ad-4e2f-b5c9-eda77578272b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks" Nov 25 17:00:00 crc kubenswrapper[4802]: I1125 17:00:00.425193 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z29mj\" (UniqueName: \"kubernetes.io/projected/2380f347-95ad-4e2f-b5c9-eda77578272b-kube-api-access-z29mj\") pod \"collect-profiles-29401500-4q2ks\" (UID: \"2380f347-95ad-4e2f-b5c9-eda77578272b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks" Nov 25 17:00:00 crc kubenswrapper[4802]: I1125 17:00:00.425254 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2380f347-95ad-4e2f-b5c9-eda77578272b-config-volume\") pod \"collect-profiles-29401500-4q2ks\" (UID: \"2380f347-95ad-4e2f-b5c9-eda77578272b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks" Nov 25 17:00:00 crc kubenswrapper[4802]: I1125 17:00:00.425278 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2380f347-95ad-4e2f-b5c9-eda77578272b-secret-volume\") pod \"collect-profiles-29401500-4q2ks\" (UID: \"2380f347-95ad-4e2f-b5c9-eda77578272b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks" Nov 25 17:00:00 crc kubenswrapper[4802]: I1125 17:00:00.426633 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2380f347-95ad-4e2f-b5c9-eda77578272b-config-volume\") pod \"collect-profiles-29401500-4q2ks\" (UID: \"2380f347-95ad-4e2f-b5c9-eda77578272b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks" Nov 25 17:00:00 crc kubenswrapper[4802]: I1125 17:00:00.434273 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2380f347-95ad-4e2f-b5c9-eda77578272b-secret-volume\") pod \"collect-profiles-29401500-4q2ks\" (UID: \"2380f347-95ad-4e2f-b5c9-eda77578272b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks" Nov 25 17:00:00 crc kubenswrapper[4802]: I1125 17:00:00.447815 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z29mj\" (UniqueName: \"kubernetes.io/projected/2380f347-95ad-4e2f-b5c9-eda77578272b-kube-api-access-z29mj\") pod \"collect-profiles-29401500-4q2ks\" (UID: \"2380f347-95ad-4e2f-b5c9-eda77578272b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks" Nov 25 17:00:00 crc kubenswrapper[4802]: I1125 17:00:00.483281 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks" Nov 25 17:00:01 crc kubenswrapper[4802]: I1125 17:00:01.908800 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks"] Nov 25 17:00:01 crc kubenswrapper[4802]: W1125 17:00:01.922903 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2380f347_95ad_4e2f_b5c9_eda77578272b.slice/crio-0e8acc9dfe2fbad6c168c9a291191942527f9e4e2dc7c4f35575129b84e3eca2 WatchSource:0}: Error finding container 0e8acc9dfe2fbad6c168c9a291191942527f9e4e2dc7c4f35575129b84e3eca2: Status 404 returned error can't find the container with id 0e8acc9dfe2fbad6c168c9a291191942527f9e4e2dc7c4f35575129b84e3eca2 Nov 25 17:00:02 crc kubenswrapper[4802]: I1125 17:00:02.602346 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-tcf29" event={"ID":"85895137-3ed8-4214-92fa-ea23980c5083","Type":"ContainerStarted","Data":"57f7851d4702365741bc79d43ed809d22c5dea10bf02cefd6186fe37e4e62ee2"} Nov 25 17:00:02 crc kubenswrapper[4802]: I1125 17:00:02.602753 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-tcf29" Nov 25 17:00:02 crc kubenswrapper[4802]: I1125 17:00:02.605884 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-xnshj" event={"ID":"cd0861df-163a-47f7-8a35-8b8c25e02ca4","Type":"ContainerStarted","Data":"107650415dbe43d3c20082b89a49f48ba1b69d127f353241f237791bf2a62691"} Nov 25 17:00:02 crc kubenswrapper[4802]: I1125 17:00:02.605985 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-xnshj" Nov 25 17:00:02 crc kubenswrapper[4802]: I1125 17:00:02.608277 4802 generic.go:334] "Generic (PLEG): container finished" podID="2380f347-95ad-4e2f-b5c9-eda77578272b" containerID="07aff647f05e4deaa3e14dccfcdbcedc4319eab96a990a4a9962c38258082abf" exitCode=0 Nov 25 17:00:02 crc kubenswrapper[4802]: I1125 17:00:02.608332 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks" event={"ID":"2380f347-95ad-4e2f-b5c9-eda77578272b","Type":"ContainerDied","Data":"07aff647f05e4deaa3e14dccfcdbcedc4319eab96a990a4a9962c38258082abf"} Nov 25 17:00:02 crc kubenswrapper[4802]: I1125 17:00:02.608365 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks" event={"ID":"2380f347-95ad-4e2f-b5c9-eda77578272b","Type":"ContainerStarted","Data":"0e8acc9dfe2fbad6c168c9a291191942527f9e4e2dc7c4f35575129b84e3eca2"} Nov 25 17:00:02 crc kubenswrapper[4802]: I1125 17:00:02.618234 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-tcf29" podStartSLOduration=2.159469829 podStartE2EDuration="5.618189424s" podCreationTimestamp="2025-11-25 16:59:57 +0000 UTC" firstStartedPulling="2025-11-25 16:59:58.068537359 +0000 UTC m=+781.212884545" lastFinishedPulling="2025-11-25 17:00:01.527256954 +0000 UTC m=+784.671604140" observedRunningTime="2025-11-25 17:00:02.617182367 +0000 UTC m=+785.761529563" watchObservedRunningTime="2025-11-25 17:00:02.618189424 +0000 UTC m=+785.762536610" Nov 25 17:00:02 crc kubenswrapper[4802]: I1125 17:00:02.663195 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-xnshj" podStartSLOduration=3.476831592 podStartE2EDuration="5.663121343s" podCreationTimestamp="2025-11-25 16:59:57 +0000 UTC" firstStartedPulling="2025-11-25 16:59:59.343765049 +0000 UTC m=+782.488112265" lastFinishedPulling="2025-11-25 17:00:01.53005483 +0000 UTC m=+784.674402016" observedRunningTime="2025-11-25 17:00:02.658793756 +0000 UTC m=+785.803140942" watchObservedRunningTime="2025-11-25 17:00:02.663121343 +0000 UTC m=+785.807468529" Nov 25 17:00:04 crc kubenswrapper[4802]: I1125 17:00:04.546051 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks" Nov 25 17:00:04 crc kubenswrapper[4802]: I1125 17:00:04.595989 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2380f347-95ad-4e2f-b5c9-eda77578272b-config-volume\") pod \"2380f347-95ad-4e2f-b5c9-eda77578272b\" (UID: \"2380f347-95ad-4e2f-b5c9-eda77578272b\") " Nov 25 17:00:04 crc kubenswrapper[4802]: I1125 17:00:04.596088 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z29mj\" (UniqueName: \"kubernetes.io/projected/2380f347-95ad-4e2f-b5c9-eda77578272b-kube-api-access-z29mj\") pod \"2380f347-95ad-4e2f-b5c9-eda77578272b\" (UID: \"2380f347-95ad-4e2f-b5c9-eda77578272b\") " Nov 25 17:00:04 crc kubenswrapper[4802]: I1125 17:00:04.596151 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2380f347-95ad-4e2f-b5c9-eda77578272b-secret-volume\") pod \"2380f347-95ad-4e2f-b5c9-eda77578272b\" (UID: \"2380f347-95ad-4e2f-b5c9-eda77578272b\") " Nov 25 17:00:04 crc kubenswrapper[4802]: I1125 17:00:04.597544 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2380f347-95ad-4e2f-b5c9-eda77578272b-config-volume" (OuterVolumeSpecName: "config-volume") pod "2380f347-95ad-4e2f-b5c9-eda77578272b" (UID: "2380f347-95ad-4e2f-b5c9-eda77578272b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:00:04 crc kubenswrapper[4802]: I1125 17:00:04.604500 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2380f347-95ad-4e2f-b5c9-eda77578272b-kube-api-access-z29mj" (OuterVolumeSpecName: "kube-api-access-z29mj") pod "2380f347-95ad-4e2f-b5c9-eda77578272b" (UID: "2380f347-95ad-4e2f-b5c9-eda77578272b"). InnerVolumeSpecName "kube-api-access-z29mj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:00:04 crc kubenswrapper[4802]: I1125 17:00:04.604649 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2380f347-95ad-4e2f-b5c9-eda77578272b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2380f347-95ad-4e2f-b5c9-eda77578272b" (UID: "2380f347-95ad-4e2f-b5c9-eda77578272b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:00:04 crc kubenswrapper[4802]: I1125 17:00:04.620700 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks" event={"ID":"2380f347-95ad-4e2f-b5c9-eda77578272b","Type":"ContainerDied","Data":"0e8acc9dfe2fbad6c168c9a291191942527f9e4e2dc7c4f35575129b84e3eca2"} Nov 25 17:00:04 crc kubenswrapper[4802]: I1125 17:00:04.620750 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401500-4q2ks" Nov 25 17:00:04 crc kubenswrapper[4802]: I1125 17:00:04.620756 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e8acc9dfe2fbad6c168c9a291191942527f9e4e2dc7c4f35575129b84e3eca2" Nov 25 17:00:04 crc kubenswrapper[4802]: I1125 17:00:04.697660 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z29mj\" (UniqueName: \"kubernetes.io/projected/2380f347-95ad-4e2f-b5c9-eda77578272b-kube-api-access-z29mj\") on node \"crc\" DevicePath \"\"" Nov 25 17:00:04 crc kubenswrapper[4802]: I1125 17:00:04.697703 4802 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2380f347-95ad-4e2f-b5c9-eda77578272b-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 17:00:04 crc kubenswrapper[4802]: I1125 17:00:04.697714 4802 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2380f347-95ad-4e2f-b5c9-eda77578272b-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 17:00:05 crc kubenswrapper[4802]: I1125 17:00:05.628397 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-rzknk" event={"ID":"d605e3c4-111d-4f35-bf63-f855001d1967","Type":"ContainerStarted","Data":"d76c889618f8b5dab9c93958fc6f52871f3b91eeeb830bc7f7465abd903439c4"} Nov 25 17:00:05 crc kubenswrapper[4802]: I1125 17:00:05.628884 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-rzknk" Nov 25 17:00:05 crc kubenswrapper[4802]: I1125 17:00:05.630317 4802 generic.go:334] "Generic (PLEG): container finished" podID="aee28310-d76a-491b-b174-e1d131d078c7" containerID="83592f0704fcb7185222f84c723387ff3bdb4b2f29fd3efece22457b084bc0b5" exitCode=0 Nov 25 17:00:05 crc kubenswrapper[4802]: I1125 17:00:05.630367 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rzq2t" event={"ID":"aee28310-d76a-491b-b174-e1d131d078c7","Type":"ContainerDied","Data":"83592f0704fcb7185222f84c723387ff3bdb4b2f29fd3efece22457b084bc0b5"} Nov 25 17:00:05 crc kubenswrapper[4802]: I1125 17:00:05.654518 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-rzknk" podStartSLOduration=1.8940768879999998 podStartE2EDuration="8.654480128s" podCreationTimestamp="2025-11-25 16:59:57 +0000 UTC" firstStartedPulling="2025-11-25 16:59:57.860431702 +0000 UTC m=+781.004778888" lastFinishedPulling="2025-11-25 17:00:04.620834942 +0000 UTC m=+787.765182128" observedRunningTime="2025-11-25 17:00:05.646511551 +0000 UTC m=+788.790858737" watchObservedRunningTime="2025-11-25 17:00:05.654480128 +0000 UTC m=+788.798827354" Nov 25 17:00:06 crc kubenswrapper[4802]: I1125 17:00:06.641093 4802 generic.go:334] "Generic (PLEG): container finished" podID="aee28310-d76a-491b-b174-e1d131d078c7" containerID="aa428497f828c2bfcc374b1f17292518d07828f807eff4060fdcaabf03283d99" exitCode=0 Nov 25 17:00:06 crc kubenswrapper[4802]: I1125 17:00:06.641199 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rzq2t" event={"ID":"aee28310-d76a-491b-b174-e1d131d078c7","Type":"ContainerDied","Data":"aa428497f828c2bfcc374b1f17292518d07828f807eff4060fdcaabf03283d99"} Nov 25 17:00:07 crc kubenswrapper[4802]: I1125 17:00:07.650142 4802 generic.go:334] "Generic (PLEG): container finished" podID="aee28310-d76a-491b-b174-e1d131d078c7" containerID="87ceaec2924dc820f7e7af1db85d63c37d7c7c9d3570eb45e1971978225fbf63" exitCode=0 Nov 25 17:00:07 crc kubenswrapper[4802]: I1125 17:00:07.650249 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rzq2t" event={"ID":"aee28310-d76a-491b-b174-e1d131d078c7","Type":"ContainerDied","Data":"87ceaec2924dc820f7e7af1db85d63c37d7c7c9d3570eb45e1971978225fbf63"} Nov 25 17:00:08 crc kubenswrapper[4802]: I1125 17:00:08.659723 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rzq2t" event={"ID":"aee28310-d76a-491b-b174-e1d131d078c7","Type":"ContainerStarted","Data":"4c6d4dc22b5fe588fc335014260f4ea5e16f178b9344a97852bda5ed5a5e4c53"} Nov 25 17:00:08 crc kubenswrapper[4802]: I1125 17:00:08.660218 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rzq2t" event={"ID":"aee28310-d76a-491b-b174-e1d131d078c7","Type":"ContainerStarted","Data":"a0b43b5495aeb2cf2800e27bc21ae5c2783fa9cd81aa052aa0a4e9712a524148"} Nov 25 17:00:08 crc kubenswrapper[4802]: I1125 17:00:08.660229 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rzq2t" event={"ID":"aee28310-d76a-491b-b174-e1d131d078c7","Type":"ContainerStarted","Data":"2c130cc53baf804bb370eba8f7d3ad867db2e5ac7853079aec74c6b9332ea7c0"} Nov 25 17:00:08 crc kubenswrapper[4802]: I1125 17:00:08.660238 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rzq2t" event={"ID":"aee28310-d76a-491b-b174-e1d131d078c7","Type":"ContainerStarted","Data":"f98a63cb9d44a7a14089d8bfa9426b0061489f4f46b891e5e9b6d33a65a2234a"} Nov 25 17:00:09 crc kubenswrapper[4802]: I1125 17:00:09.023596 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-xnshj" Nov 25 17:00:09 crc kubenswrapper[4802]: I1125 17:00:09.671424 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rzq2t" event={"ID":"aee28310-d76a-491b-b174-e1d131d078c7","Type":"ContainerStarted","Data":"6f34f7f8e96bce7de80e59892ebea800b6a2f87b85eb154ba2e9e8ec308eba79"} Nov 25 17:00:09 crc kubenswrapper[4802]: I1125 17:00:09.671888 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-rzq2t" Nov 25 17:00:09 crc kubenswrapper[4802]: I1125 17:00:09.671903 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rzq2t" event={"ID":"aee28310-d76a-491b-b174-e1d131d078c7","Type":"ContainerStarted","Data":"70cea515c8a1b04184b2593d41e2d840106eecbc9696979b4938d1aebe639866"} Nov 25 17:00:09 crc kubenswrapper[4802]: I1125 17:00:09.694568 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-rzq2t" podStartSLOduration=5.663110372 podStartE2EDuration="12.694540446s" podCreationTimestamp="2025-11-25 16:59:57 +0000 UTC" firstStartedPulling="2025-11-25 16:59:57.578450151 +0000 UTC m=+780.722797337" lastFinishedPulling="2025-11-25 17:00:04.609880225 +0000 UTC m=+787.754227411" observedRunningTime="2025-11-25 17:00:09.69134747 +0000 UTC m=+792.835694666" watchObservedRunningTime="2025-11-25 17:00:09.694540446 +0000 UTC m=+792.838887632" Nov 25 17:00:12 crc kubenswrapper[4802]: I1125 17:00:12.416631 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-rzq2t" Nov 25 17:00:12 crc kubenswrapper[4802]: I1125 17:00:12.453984 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-rzq2t" Nov 25 17:00:12 crc kubenswrapper[4802]: I1125 17:00:12.504724 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8df47"] Nov 25 17:00:12 crc kubenswrapper[4802]: E1125 17:00:12.512969 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2380f347-95ad-4e2f-b5c9-eda77578272b" containerName="collect-profiles" Nov 25 17:00:12 crc kubenswrapper[4802]: I1125 17:00:12.513020 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="2380f347-95ad-4e2f-b5c9-eda77578272b" containerName="collect-profiles" Nov 25 17:00:12 crc kubenswrapper[4802]: I1125 17:00:12.513504 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="2380f347-95ad-4e2f-b5c9-eda77578272b" containerName="collect-profiles" Nov 25 17:00:12 crc kubenswrapper[4802]: I1125 17:00:12.516479 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8df47" Nov 25 17:00:12 crc kubenswrapper[4802]: I1125 17:00:12.527009 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8df47"] Nov 25 17:00:12 crc kubenswrapper[4802]: I1125 17:00:12.610288 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3449cbd1-efbe-4a67-ada6-204c7119b027-catalog-content\") pod \"community-operators-8df47\" (UID: \"3449cbd1-efbe-4a67-ada6-204c7119b027\") " pod="openshift-marketplace/community-operators-8df47" Nov 25 17:00:12 crc kubenswrapper[4802]: I1125 17:00:12.610731 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbgbl\" (UniqueName: \"kubernetes.io/projected/3449cbd1-efbe-4a67-ada6-204c7119b027-kube-api-access-jbgbl\") pod \"community-operators-8df47\" (UID: \"3449cbd1-efbe-4a67-ada6-204c7119b027\") " pod="openshift-marketplace/community-operators-8df47" Nov 25 17:00:12 crc kubenswrapper[4802]: I1125 17:00:12.610939 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3449cbd1-efbe-4a67-ada6-204c7119b027-utilities\") pod \"community-operators-8df47\" (UID: \"3449cbd1-efbe-4a67-ada6-204c7119b027\") " pod="openshift-marketplace/community-operators-8df47" Nov 25 17:00:12 crc kubenswrapper[4802]: I1125 17:00:12.711810 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3449cbd1-efbe-4a67-ada6-204c7119b027-catalog-content\") pod \"community-operators-8df47\" (UID: \"3449cbd1-efbe-4a67-ada6-204c7119b027\") " pod="openshift-marketplace/community-operators-8df47" Nov 25 17:00:12 crc kubenswrapper[4802]: I1125 17:00:12.711934 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbgbl\" (UniqueName: \"kubernetes.io/projected/3449cbd1-efbe-4a67-ada6-204c7119b027-kube-api-access-jbgbl\") pod \"community-operators-8df47\" (UID: \"3449cbd1-efbe-4a67-ada6-204c7119b027\") " pod="openshift-marketplace/community-operators-8df47" Nov 25 17:00:12 crc kubenswrapper[4802]: I1125 17:00:12.712013 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3449cbd1-efbe-4a67-ada6-204c7119b027-utilities\") pod \"community-operators-8df47\" (UID: \"3449cbd1-efbe-4a67-ada6-204c7119b027\") " pod="openshift-marketplace/community-operators-8df47" Nov 25 17:00:12 crc kubenswrapper[4802]: I1125 17:00:12.712687 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3449cbd1-efbe-4a67-ada6-204c7119b027-utilities\") pod \"community-operators-8df47\" (UID: \"3449cbd1-efbe-4a67-ada6-204c7119b027\") " pod="openshift-marketplace/community-operators-8df47" Nov 25 17:00:12 crc kubenswrapper[4802]: I1125 17:00:12.712800 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3449cbd1-efbe-4a67-ada6-204c7119b027-catalog-content\") pod \"community-operators-8df47\" (UID: \"3449cbd1-efbe-4a67-ada6-204c7119b027\") " pod="openshift-marketplace/community-operators-8df47" Nov 25 17:00:12 crc kubenswrapper[4802]: I1125 17:00:12.738423 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbgbl\" (UniqueName: \"kubernetes.io/projected/3449cbd1-efbe-4a67-ada6-204c7119b027-kube-api-access-jbgbl\") pod \"community-operators-8df47\" (UID: \"3449cbd1-efbe-4a67-ada6-204c7119b027\") " pod="openshift-marketplace/community-operators-8df47" Nov 25 17:00:12 crc kubenswrapper[4802]: I1125 17:00:12.844949 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8df47" Nov 25 17:00:13 crc kubenswrapper[4802]: I1125 17:00:13.152258 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8df47"] Nov 25 17:00:13 crc kubenswrapper[4802]: W1125 17:00:13.158635 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3449cbd1_efbe_4a67_ada6_204c7119b027.slice/crio-583bff679662eca29642def5a9fcc4ea0f113b0b07daefa61f3c0ceab80878e2 WatchSource:0}: Error finding container 583bff679662eca29642def5a9fcc4ea0f113b0b07daefa61f3c0ceab80878e2: Status 404 returned error can't find the container with id 583bff679662eca29642def5a9fcc4ea0f113b0b07daefa61f3c0ceab80878e2 Nov 25 17:00:13 crc kubenswrapper[4802]: I1125 17:00:13.699609 4802 generic.go:334] "Generic (PLEG): container finished" podID="3449cbd1-efbe-4a67-ada6-204c7119b027" containerID="68c7c689b59142dfefa5a658666b0a497a0ace7e8e7aac8da1b68b62864674c1" exitCode=0 Nov 25 17:00:13 crc kubenswrapper[4802]: I1125 17:00:13.699662 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8df47" event={"ID":"3449cbd1-efbe-4a67-ada6-204c7119b027","Type":"ContainerDied","Data":"68c7c689b59142dfefa5a658666b0a497a0ace7e8e7aac8da1b68b62864674c1"} Nov 25 17:00:13 crc kubenswrapper[4802]: I1125 17:00:13.699701 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8df47" event={"ID":"3449cbd1-efbe-4a67-ada6-204c7119b027","Type":"ContainerStarted","Data":"583bff679662eca29642def5a9fcc4ea0f113b0b07daefa61f3c0ceab80878e2"} Nov 25 17:00:14 crc kubenswrapper[4802]: I1125 17:00:14.710304 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8df47" event={"ID":"3449cbd1-efbe-4a67-ada6-204c7119b027","Type":"ContainerStarted","Data":"f6b4aaab1fa762becc3f4049a768d743f564e7b2982901513fce82bcc9140e6b"} Nov 25 17:00:15 crc kubenswrapper[4802]: I1125 17:00:15.718712 4802 generic.go:334] "Generic (PLEG): container finished" podID="3449cbd1-efbe-4a67-ada6-204c7119b027" containerID="f6b4aaab1fa762becc3f4049a768d743f564e7b2982901513fce82bcc9140e6b" exitCode=0 Nov 25 17:00:15 crc kubenswrapper[4802]: I1125 17:00:15.718773 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8df47" event={"ID":"3449cbd1-efbe-4a67-ada6-204c7119b027","Type":"ContainerDied","Data":"f6b4aaab1fa762becc3f4049a768d743f564e7b2982901513fce82bcc9140e6b"} Nov 25 17:00:16 crc kubenswrapper[4802]: I1125 17:00:16.729910 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8df47" event={"ID":"3449cbd1-efbe-4a67-ada6-204c7119b027","Type":"ContainerStarted","Data":"3fe23e0100d7b3078a4752a6efa9a40c214031de714ed08a16f2ae8c5c23b62c"} Nov 25 17:00:16 crc kubenswrapper[4802]: I1125 17:00:16.751100 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8df47" podStartSLOduration=2.224066445 podStartE2EDuration="4.751079651s" podCreationTimestamp="2025-11-25 17:00:12 +0000 UTC" firstStartedPulling="2025-11-25 17:00:13.701479276 +0000 UTC m=+796.845826462" lastFinishedPulling="2025-11-25 17:00:16.228492492 +0000 UTC m=+799.372839668" observedRunningTime="2025-11-25 17:00:16.749437257 +0000 UTC m=+799.893784443" watchObservedRunningTime="2025-11-25 17:00:16.751079651 +0000 UTC m=+799.895426837" Nov 25 17:00:17 crc kubenswrapper[4802]: I1125 17:00:17.418897 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-rzq2t" Nov 25 17:00:17 crc kubenswrapper[4802]: I1125 17:00:17.472879 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-rzknk" Nov 25 17:00:17 crc kubenswrapper[4802]: I1125 17:00:17.530907 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-tcf29" Nov 25 17:00:19 crc kubenswrapper[4802]: I1125 17:00:19.696474 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-nwwvv"] Nov 25 17:00:19 crc kubenswrapper[4802]: I1125 17:00:19.697675 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-nwwvv" Nov 25 17:00:19 crc kubenswrapper[4802]: I1125 17:00:19.700546 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-dllpk" Nov 25 17:00:19 crc kubenswrapper[4802]: I1125 17:00:19.700884 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 25 17:00:19 crc kubenswrapper[4802]: I1125 17:00:19.702848 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 25 17:00:19 crc kubenswrapper[4802]: I1125 17:00:19.711456 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-nwwvv"] Nov 25 17:00:19 crc kubenswrapper[4802]: I1125 17:00:19.713867 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56wlg\" (UniqueName: \"kubernetes.io/projected/ad830c47-f0b5-48d7-874f-41f78fbaf23d-kube-api-access-56wlg\") pod \"mariadb-operator-index-nwwvv\" (UID: \"ad830c47-f0b5-48d7-874f-41f78fbaf23d\") " pod="openstack-operators/mariadb-operator-index-nwwvv" Nov 25 17:00:19 crc kubenswrapper[4802]: I1125 17:00:19.814962 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56wlg\" (UniqueName: \"kubernetes.io/projected/ad830c47-f0b5-48d7-874f-41f78fbaf23d-kube-api-access-56wlg\") pod \"mariadb-operator-index-nwwvv\" (UID: \"ad830c47-f0b5-48d7-874f-41f78fbaf23d\") " pod="openstack-operators/mariadb-operator-index-nwwvv" Nov 25 17:00:19 crc kubenswrapper[4802]: I1125 17:00:19.836914 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56wlg\" (UniqueName: \"kubernetes.io/projected/ad830c47-f0b5-48d7-874f-41f78fbaf23d-kube-api-access-56wlg\") pod \"mariadb-operator-index-nwwvv\" (UID: \"ad830c47-f0b5-48d7-874f-41f78fbaf23d\") " pod="openstack-operators/mariadb-operator-index-nwwvv" Nov 25 17:00:20 crc kubenswrapper[4802]: I1125 17:00:20.016234 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-nwwvv" Nov 25 17:00:20 crc kubenswrapper[4802]: I1125 17:00:20.307689 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-nwwvv"] Nov 25 17:00:20 crc kubenswrapper[4802]: W1125 17:00:20.314427 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad830c47_f0b5_48d7_874f_41f78fbaf23d.slice/crio-324632c7191cf01b9658512027ebcccb9de75ccfa1cf9246ccaaed4c6101af07 WatchSource:0}: Error finding container 324632c7191cf01b9658512027ebcccb9de75ccfa1cf9246ccaaed4c6101af07: Status 404 returned error can't find the container with id 324632c7191cf01b9658512027ebcccb9de75ccfa1cf9246ccaaed4c6101af07 Nov 25 17:00:20 crc kubenswrapper[4802]: I1125 17:00:20.845531 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-nwwvv" event={"ID":"ad830c47-f0b5-48d7-874f-41f78fbaf23d","Type":"ContainerStarted","Data":"324632c7191cf01b9658512027ebcccb9de75ccfa1cf9246ccaaed4c6101af07"} Nov 25 17:00:21 crc kubenswrapper[4802]: I1125 17:00:21.853328 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-nwwvv" event={"ID":"ad830c47-f0b5-48d7-874f-41f78fbaf23d","Type":"ContainerStarted","Data":"fc1a8a2db800b9ed413728aea851629de6d4f01bc614a1767f743fccc78ca7a0"} Nov 25 17:00:21 crc kubenswrapper[4802]: I1125 17:00:21.869012 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-nwwvv" podStartSLOduration=2.115148132 podStartE2EDuration="2.868981815s" podCreationTimestamp="2025-11-25 17:00:19 +0000 UTC" firstStartedPulling="2025-11-25 17:00:20.316652387 +0000 UTC m=+803.460999573" lastFinishedPulling="2025-11-25 17:00:21.07048607 +0000 UTC m=+804.214833256" observedRunningTime="2025-11-25 17:00:21.866105697 +0000 UTC m=+805.010452883" watchObservedRunningTime="2025-11-25 17:00:21.868981815 +0000 UTC m=+805.013329001" Nov 25 17:00:22 crc kubenswrapper[4802]: I1125 17:00:22.845504 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8df47" Nov 25 17:00:22 crc kubenswrapper[4802]: I1125 17:00:22.845906 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8df47" Nov 25 17:00:22 crc kubenswrapper[4802]: I1125 17:00:22.901237 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8df47" Nov 25 17:00:22 crc kubenswrapper[4802]: I1125 17:00:22.945908 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8df47" Nov 25 17:00:24 crc kubenswrapper[4802]: I1125 17:00:24.089477 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8df47"] Nov 25 17:00:24 crc kubenswrapper[4802]: I1125 17:00:24.872608 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8df47" podUID="3449cbd1-efbe-4a67-ada6-204c7119b027" containerName="registry-server" containerID="cri-o://3fe23e0100d7b3078a4752a6efa9a40c214031de714ed08a16f2ae8c5c23b62c" gracePeriod=2 Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.314470 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8df47" Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.499396 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3449cbd1-efbe-4a67-ada6-204c7119b027-catalog-content\") pod \"3449cbd1-efbe-4a67-ada6-204c7119b027\" (UID: \"3449cbd1-efbe-4a67-ada6-204c7119b027\") " Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.499640 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3449cbd1-efbe-4a67-ada6-204c7119b027-utilities\") pod \"3449cbd1-efbe-4a67-ada6-204c7119b027\" (UID: \"3449cbd1-efbe-4a67-ada6-204c7119b027\") " Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.499712 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbgbl\" (UniqueName: \"kubernetes.io/projected/3449cbd1-efbe-4a67-ada6-204c7119b027-kube-api-access-jbgbl\") pod \"3449cbd1-efbe-4a67-ada6-204c7119b027\" (UID: \"3449cbd1-efbe-4a67-ada6-204c7119b027\") " Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.501271 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3449cbd1-efbe-4a67-ada6-204c7119b027-utilities" (OuterVolumeSpecName: "utilities") pod "3449cbd1-efbe-4a67-ada6-204c7119b027" (UID: "3449cbd1-efbe-4a67-ada6-204c7119b027"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.509714 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3449cbd1-efbe-4a67-ada6-204c7119b027-kube-api-access-jbgbl" (OuterVolumeSpecName: "kube-api-access-jbgbl") pod "3449cbd1-efbe-4a67-ada6-204c7119b027" (UID: "3449cbd1-efbe-4a67-ada6-204c7119b027"). InnerVolumeSpecName "kube-api-access-jbgbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.584041 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3449cbd1-efbe-4a67-ada6-204c7119b027-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3449cbd1-efbe-4a67-ada6-204c7119b027" (UID: "3449cbd1-efbe-4a67-ada6-204c7119b027"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.602088 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3449cbd1-efbe-4a67-ada6-204c7119b027-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.602146 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbgbl\" (UniqueName: \"kubernetes.io/projected/3449cbd1-efbe-4a67-ada6-204c7119b027-kube-api-access-jbgbl\") on node \"crc\" DevicePath \"\"" Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.602159 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3449cbd1-efbe-4a67-ada6-204c7119b027-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.881519 4802 generic.go:334] "Generic (PLEG): container finished" podID="3449cbd1-efbe-4a67-ada6-204c7119b027" containerID="3fe23e0100d7b3078a4752a6efa9a40c214031de714ed08a16f2ae8c5c23b62c" exitCode=0 Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.881586 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8df47" event={"ID":"3449cbd1-efbe-4a67-ada6-204c7119b027","Type":"ContainerDied","Data":"3fe23e0100d7b3078a4752a6efa9a40c214031de714ed08a16f2ae8c5c23b62c"} Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.881625 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8df47" Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.881652 4802 scope.go:117] "RemoveContainer" containerID="3fe23e0100d7b3078a4752a6efa9a40c214031de714ed08a16f2ae8c5c23b62c" Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.881633 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8df47" event={"ID":"3449cbd1-efbe-4a67-ada6-204c7119b027","Type":"ContainerDied","Data":"583bff679662eca29642def5a9fcc4ea0f113b0b07daefa61f3c0ceab80878e2"} Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.904951 4802 scope.go:117] "RemoveContainer" containerID="f6b4aaab1fa762becc3f4049a768d743f564e7b2982901513fce82bcc9140e6b" Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.910809 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8df47"] Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.917350 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8df47"] Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.943461 4802 scope.go:117] "RemoveContainer" containerID="68c7c689b59142dfefa5a658666b0a497a0ace7e8e7aac8da1b68b62864674c1" Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.967730 4802 scope.go:117] "RemoveContainer" containerID="3fe23e0100d7b3078a4752a6efa9a40c214031de714ed08a16f2ae8c5c23b62c" Nov 25 17:00:25 crc kubenswrapper[4802]: E1125 17:00:25.968199 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fe23e0100d7b3078a4752a6efa9a40c214031de714ed08a16f2ae8c5c23b62c\": container with ID starting with 3fe23e0100d7b3078a4752a6efa9a40c214031de714ed08a16f2ae8c5c23b62c not found: ID does not exist" containerID="3fe23e0100d7b3078a4752a6efa9a40c214031de714ed08a16f2ae8c5c23b62c" Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.968245 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fe23e0100d7b3078a4752a6efa9a40c214031de714ed08a16f2ae8c5c23b62c"} err="failed to get container status \"3fe23e0100d7b3078a4752a6efa9a40c214031de714ed08a16f2ae8c5c23b62c\": rpc error: code = NotFound desc = could not find container \"3fe23e0100d7b3078a4752a6efa9a40c214031de714ed08a16f2ae8c5c23b62c\": container with ID starting with 3fe23e0100d7b3078a4752a6efa9a40c214031de714ed08a16f2ae8c5c23b62c not found: ID does not exist" Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.968294 4802 scope.go:117] "RemoveContainer" containerID="f6b4aaab1fa762becc3f4049a768d743f564e7b2982901513fce82bcc9140e6b" Nov 25 17:00:25 crc kubenswrapper[4802]: E1125 17:00:25.968538 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6b4aaab1fa762becc3f4049a768d743f564e7b2982901513fce82bcc9140e6b\": container with ID starting with f6b4aaab1fa762becc3f4049a768d743f564e7b2982901513fce82bcc9140e6b not found: ID does not exist" containerID="f6b4aaab1fa762becc3f4049a768d743f564e7b2982901513fce82bcc9140e6b" Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.968566 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6b4aaab1fa762becc3f4049a768d743f564e7b2982901513fce82bcc9140e6b"} err="failed to get container status \"f6b4aaab1fa762becc3f4049a768d743f564e7b2982901513fce82bcc9140e6b\": rpc error: code = NotFound desc = could not find container \"f6b4aaab1fa762becc3f4049a768d743f564e7b2982901513fce82bcc9140e6b\": container with ID starting with f6b4aaab1fa762becc3f4049a768d743f564e7b2982901513fce82bcc9140e6b not found: ID does not exist" Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.968581 4802 scope.go:117] "RemoveContainer" containerID="68c7c689b59142dfefa5a658666b0a497a0ace7e8e7aac8da1b68b62864674c1" Nov 25 17:00:25 crc kubenswrapper[4802]: E1125 17:00:25.968799 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68c7c689b59142dfefa5a658666b0a497a0ace7e8e7aac8da1b68b62864674c1\": container with ID starting with 68c7c689b59142dfefa5a658666b0a497a0ace7e8e7aac8da1b68b62864674c1 not found: ID does not exist" containerID="68c7c689b59142dfefa5a658666b0a497a0ace7e8e7aac8da1b68b62864674c1" Nov 25 17:00:25 crc kubenswrapper[4802]: I1125 17:00:25.968821 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68c7c689b59142dfefa5a658666b0a497a0ace7e8e7aac8da1b68b62864674c1"} err="failed to get container status \"68c7c689b59142dfefa5a658666b0a497a0ace7e8e7aac8da1b68b62864674c1\": rpc error: code = NotFound desc = could not find container \"68c7c689b59142dfefa5a658666b0a497a0ace7e8e7aac8da1b68b62864674c1\": container with ID starting with 68c7c689b59142dfefa5a658666b0a497a0ace7e8e7aac8da1b68b62864674c1 not found: ID does not exist" Nov 25 17:00:27 crc kubenswrapper[4802]: I1125 17:00:27.513566 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3449cbd1-efbe-4a67-ada6-204c7119b027" path="/var/lib/kubelet/pods/3449cbd1-efbe-4a67-ada6-204c7119b027/volumes" Nov 25 17:00:30 crc kubenswrapper[4802]: I1125 17:00:30.017522 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-nwwvv" Nov 25 17:00:30 crc kubenswrapper[4802]: I1125 17:00:30.017922 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-nwwvv" Nov 25 17:00:30 crc kubenswrapper[4802]: I1125 17:00:30.046858 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-nwwvv" Nov 25 17:00:30 crc kubenswrapper[4802]: I1125 17:00:30.938784 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-nwwvv" Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.144011 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc"] Nov 25 17:00:33 crc kubenswrapper[4802]: E1125 17:00:33.144353 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3449cbd1-efbe-4a67-ada6-204c7119b027" containerName="extract-content" Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.144368 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="3449cbd1-efbe-4a67-ada6-204c7119b027" containerName="extract-content" Nov 25 17:00:33 crc kubenswrapper[4802]: E1125 17:00:33.144389 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3449cbd1-efbe-4a67-ada6-204c7119b027" containerName="extract-utilities" Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.144397 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="3449cbd1-efbe-4a67-ada6-204c7119b027" containerName="extract-utilities" Nov 25 17:00:33 crc kubenswrapper[4802]: E1125 17:00:33.144416 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3449cbd1-efbe-4a67-ada6-204c7119b027" containerName="registry-server" Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.144426 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="3449cbd1-efbe-4a67-ada6-204c7119b027" containerName="registry-server" Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.144584 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="3449cbd1-efbe-4a67-ada6-204c7119b027" containerName="registry-server" Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.145607 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.147895 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-t7662" Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.219211 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc"] Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.305477 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d3f9606c-d757-40c8-a32e-52b5d9572f65-util\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc\" (UID: \"d3f9606c-d757-40c8-a32e-52b5d9572f65\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.305818 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjrwv\" (UniqueName: \"kubernetes.io/projected/d3f9606c-d757-40c8-a32e-52b5d9572f65-kube-api-access-rjrwv\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc\" (UID: \"d3f9606c-d757-40c8-a32e-52b5d9572f65\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.305941 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d3f9606c-d757-40c8-a32e-52b5d9572f65-bundle\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc\" (UID: \"d3f9606c-d757-40c8-a32e-52b5d9572f65\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.406645 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d3f9606c-d757-40c8-a32e-52b5d9572f65-util\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc\" (UID: \"d3f9606c-d757-40c8-a32e-52b5d9572f65\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.406701 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjrwv\" (UniqueName: \"kubernetes.io/projected/d3f9606c-d757-40c8-a32e-52b5d9572f65-kube-api-access-rjrwv\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc\" (UID: \"d3f9606c-d757-40c8-a32e-52b5d9572f65\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.406738 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d3f9606c-d757-40c8-a32e-52b5d9572f65-bundle\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc\" (UID: \"d3f9606c-d757-40c8-a32e-52b5d9572f65\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.407307 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d3f9606c-d757-40c8-a32e-52b5d9572f65-util\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc\" (UID: \"d3f9606c-d757-40c8-a32e-52b5d9572f65\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.407323 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d3f9606c-d757-40c8-a32e-52b5d9572f65-bundle\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc\" (UID: \"d3f9606c-d757-40c8-a32e-52b5d9572f65\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.425062 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjrwv\" (UniqueName: \"kubernetes.io/projected/d3f9606c-d757-40c8-a32e-52b5d9572f65-kube-api-access-rjrwv\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc\" (UID: \"d3f9606c-d757-40c8-a32e-52b5d9572f65\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.463929 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.897701 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc"] Nov 25 17:00:33 crc kubenswrapper[4802]: I1125 17:00:33.935601 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" event={"ID":"d3f9606c-d757-40c8-a32e-52b5d9572f65","Type":"ContainerStarted","Data":"7ba475c1dbd5e469700bddd9b20f7046eee733239d2cf4034c2dc7544dbd2b03"} Nov 25 17:00:34 crc kubenswrapper[4802]: I1125 17:00:34.947574 4802 generic.go:334] "Generic (PLEG): container finished" podID="d3f9606c-d757-40c8-a32e-52b5d9572f65" containerID="c3e97add8eaab05dcf91b2b1169d3cd2d6141c1aac7c7a8070bd2da0283e6fdc" exitCode=0 Nov 25 17:00:34 crc kubenswrapper[4802]: I1125 17:00:34.947660 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" event={"ID":"d3f9606c-d757-40c8-a32e-52b5d9572f65","Type":"ContainerDied","Data":"c3e97add8eaab05dcf91b2b1169d3cd2d6141c1aac7c7a8070bd2da0283e6fdc"} Nov 25 17:00:35 crc kubenswrapper[4802]: I1125 17:00:35.955077 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" event={"ID":"d3f9606c-d757-40c8-a32e-52b5d9572f65","Type":"ContainerStarted","Data":"1a1631a56e3fecd68036c4a0c758ef189fbc02c1d10b8598cc297e9cdf8e87d5"} Nov 25 17:00:36 crc kubenswrapper[4802]: I1125 17:00:36.094929 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kcw4t"] Nov 25 17:00:36 crc kubenswrapper[4802]: I1125 17:00:36.096627 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kcw4t" Nov 25 17:00:36 crc kubenswrapper[4802]: I1125 17:00:36.106367 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kcw4t"] Nov 25 17:00:36 crc kubenswrapper[4802]: I1125 17:00:36.246048 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97fc5986-61ea-49b8-a339-cd4526610431-utilities\") pod \"redhat-marketplace-kcw4t\" (UID: \"97fc5986-61ea-49b8-a339-cd4526610431\") " pod="openshift-marketplace/redhat-marketplace-kcw4t" Nov 25 17:00:36 crc kubenswrapper[4802]: I1125 17:00:36.246107 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97fc5986-61ea-49b8-a339-cd4526610431-catalog-content\") pod \"redhat-marketplace-kcw4t\" (UID: \"97fc5986-61ea-49b8-a339-cd4526610431\") " pod="openshift-marketplace/redhat-marketplace-kcw4t" Nov 25 17:00:36 crc kubenswrapper[4802]: I1125 17:00:36.246161 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9smdc\" (UniqueName: \"kubernetes.io/projected/97fc5986-61ea-49b8-a339-cd4526610431-kube-api-access-9smdc\") pod \"redhat-marketplace-kcw4t\" (UID: \"97fc5986-61ea-49b8-a339-cd4526610431\") " pod="openshift-marketplace/redhat-marketplace-kcw4t" Nov 25 17:00:36 crc kubenswrapper[4802]: I1125 17:00:36.347961 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97fc5986-61ea-49b8-a339-cd4526610431-utilities\") pod \"redhat-marketplace-kcw4t\" (UID: \"97fc5986-61ea-49b8-a339-cd4526610431\") " pod="openshift-marketplace/redhat-marketplace-kcw4t" Nov 25 17:00:36 crc kubenswrapper[4802]: I1125 17:00:36.348025 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97fc5986-61ea-49b8-a339-cd4526610431-catalog-content\") pod \"redhat-marketplace-kcw4t\" (UID: \"97fc5986-61ea-49b8-a339-cd4526610431\") " pod="openshift-marketplace/redhat-marketplace-kcw4t" Nov 25 17:00:36 crc kubenswrapper[4802]: I1125 17:00:36.348062 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9smdc\" (UniqueName: \"kubernetes.io/projected/97fc5986-61ea-49b8-a339-cd4526610431-kube-api-access-9smdc\") pod \"redhat-marketplace-kcw4t\" (UID: \"97fc5986-61ea-49b8-a339-cd4526610431\") " pod="openshift-marketplace/redhat-marketplace-kcw4t" Nov 25 17:00:36 crc kubenswrapper[4802]: I1125 17:00:36.348608 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97fc5986-61ea-49b8-a339-cd4526610431-catalog-content\") pod \"redhat-marketplace-kcw4t\" (UID: \"97fc5986-61ea-49b8-a339-cd4526610431\") " pod="openshift-marketplace/redhat-marketplace-kcw4t" Nov 25 17:00:36 crc kubenswrapper[4802]: I1125 17:00:36.348618 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97fc5986-61ea-49b8-a339-cd4526610431-utilities\") pod \"redhat-marketplace-kcw4t\" (UID: \"97fc5986-61ea-49b8-a339-cd4526610431\") " pod="openshift-marketplace/redhat-marketplace-kcw4t" Nov 25 17:00:36 crc kubenswrapper[4802]: I1125 17:00:36.367316 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9smdc\" (UniqueName: \"kubernetes.io/projected/97fc5986-61ea-49b8-a339-cd4526610431-kube-api-access-9smdc\") pod \"redhat-marketplace-kcw4t\" (UID: \"97fc5986-61ea-49b8-a339-cd4526610431\") " pod="openshift-marketplace/redhat-marketplace-kcw4t" Nov 25 17:00:36 crc kubenswrapper[4802]: I1125 17:00:36.413634 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kcw4t" Nov 25 17:00:36 crc kubenswrapper[4802]: I1125 17:00:36.834134 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kcw4t"] Nov 25 17:00:36 crc kubenswrapper[4802]: I1125 17:00:36.962707 4802 generic.go:334] "Generic (PLEG): container finished" podID="d3f9606c-d757-40c8-a32e-52b5d9572f65" containerID="1a1631a56e3fecd68036c4a0c758ef189fbc02c1d10b8598cc297e9cdf8e87d5" exitCode=0 Nov 25 17:00:36 crc kubenswrapper[4802]: I1125 17:00:36.962782 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" event={"ID":"d3f9606c-d757-40c8-a32e-52b5d9572f65","Type":"ContainerDied","Data":"1a1631a56e3fecd68036c4a0c758ef189fbc02c1d10b8598cc297e9cdf8e87d5"} Nov 25 17:00:36 crc kubenswrapper[4802]: I1125 17:00:36.965609 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kcw4t" event={"ID":"97fc5986-61ea-49b8-a339-cd4526610431","Type":"ContainerStarted","Data":"09ebfe05736cca36f6abdd1b3d81b71f64ff53bd80c02c4f02ce226f96a6779f"} Nov 25 17:00:37 crc kubenswrapper[4802]: I1125 17:00:37.294722 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-khtpq"] Nov 25 17:00:37 crc kubenswrapper[4802]: I1125 17:00:37.296787 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-khtpq" Nov 25 17:00:37 crc kubenswrapper[4802]: I1125 17:00:37.307608 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-khtpq"] Nov 25 17:00:37 crc kubenswrapper[4802]: I1125 17:00:37.362335 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/015fc770-4272-466e-99b9-871f5bfd5c79-utilities\") pod \"certified-operators-khtpq\" (UID: \"015fc770-4272-466e-99b9-871f5bfd5c79\") " pod="openshift-marketplace/certified-operators-khtpq" Nov 25 17:00:37 crc kubenswrapper[4802]: I1125 17:00:37.362461 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrpk9\" (UniqueName: \"kubernetes.io/projected/015fc770-4272-466e-99b9-871f5bfd5c79-kube-api-access-zrpk9\") pod \"certified-operators-khtpq\" (UID: \"015fc770-4272-466e-99b9-871f5bfd5c79\") " pod="openshift-marketplace/certified-operators-khtpq" Nov 25 17:00:37 crc kubenswrapper[4802]: I1125 17:00:37.362553 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/015fc770-4272-466e-99b9-871f5bfd5c79-catalog-content\") pod \"certified-operators-khtpq\" (UID: \"015fc770-4272-466e-99b9-871f5bfd5c79\") " pod="openshift-marketplace/certified-operators-khtpq" Nov 25 17:00:37 crc kubenswrapper[4802]: I1125 17:00:37.465333 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/015fc770-4272-466e-99b9-871f5bfd5c79-catalog-content\") pod \"certified-operators-khtpq\" (UID: \"015fc770-4272-466e-99b9-871f5bfd5c79\") " pod="openshift-marketplace/certified-operators-khtpq" Nov 25 17:00:37 crc kubenswrapper[4802]: I1125 17:00:37.465419 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/015fc770-4272-466e-99b9-871f5bfd5c79-utilities\") pod \"certified-operators-khtpq\" (UID: \"015fc770-4272-466e-99b9-871f5bfd5c79\") " pod="openshift-marketplace/certified-operators-khtpq" Nov 25 17:00:37 crc kubenswrapper[4802]: I1125 17:00:37.465475 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrpk9\" (UniqueName: \"kubernetes.io/projected/015fc770-4272-466e-99b9-871f5bfd5c79-kube-api-access-zrpk9\") pod \"certified-operators-khtpq\" (UID: \"015fc770-4272-466e-99b9-871f5bfd5c79\") " pod="openshift-marketplace/certified-operators-khtpq" Nov 25 17:00:37 crc kubenswrapper[4802]: I1125 17:00:37.466495 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/015fc770-4272-466e-99b9-871f5bfd5c79-catalog-content\") pod \"certified-operators-khtpq\" (UID: \"015fc770-4272-466e-99b9-871f5bfd5c79\") " pod="openshift-marketplace/certified-operators-khtpq" Nov 25 17:00:37 crc kubenswrapper[4802]: I1125 17:00:37.466935 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/015fc770-4272-466e-99b9-871f5bfd5c79-utilities\") pod \"certified-operators-khtpq\" (UID: \"015fc770-4272-466e-99b9-871f5bfd5c79\") " pod="openshift-marketplace/certified-operators-khtpq" Nov 25 17:00:37 crc kubenswrapper[4802]: I1125 17:00:37.488939 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrpk9\" (UniqueName: \"kubernetes.io/projected/015fc770-4272-466e-99b9-871f5bfd5c79-kube-api-access-zrpk9\") pod \"certified-operators-khtpq\" (UID: \"015fc770-4272-466e-99b9-871f5bfd5c79\") " pod="openshift-marketplace/certified-operators-khtpq" Nov 25 17:00:37 crc kubenswrapper[4802]: I1125 17:00:37.614559 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-khtpq" Nov 25 17:00:37 crc kubenswrapper[4802]: I1125 17:00:37.972339 4802 generic.go:334] "Generic (PLEG): container finished" podID="97fc5986-61ea-49b8-a339-cd4526610431" containerID="154d8d9cd3d99be5e2246b3befcf5bb4460dd8a6e57aef574b90170c61ee48a1" exitCode=0 Nov 25 17:00:37 crc kubenswrapper[4802]: I1125 17:00:37.972436 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kcw4t" event={"ID":"97fc5986-61ea-49b8-a339-cd4526610431","Type":"ContainerDied","Data":"154d8d9cd3d99be5e2246b3befcf5bb4460dd8a6e57aef574b90170c61ee48a1"} Nov 25 17:00:37 crc kubenswrapper[4802]: I1125 17:00:37.976784 4802 generic.go:334] "Generic (PLEG): container finished" podID="d3f9606c-d757-40c8-a32e-52b5d9572f65" containerID="f2a4a9bf1203260150f579980e251599ee94b90afee0839f11c26e03c5783e7b" exitCode=0 Nov 25 17:00:37 crc kubenswrapper[4802]: I1125 17:00:37.976874 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" event={"ID":"d3f9606c-d757-40c8-a32e-52b5d9572f65","Type":"ContainerDied","Data":"f2a4a9bf1203260150f579980e251599ee94b90afee0839f11c26e03c5783e7b"} Nov 25 17:00:38 crc kubenswrapper[4802]: I1125 17:00:38.038890 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-khtpq"] Nov 25 17:00:38 crc kubenswrapper[4802]: I1125 17:00:38.985612 4802 generic.go:334] "Generic (PLEG): container finished" podID="015fc770-4272-466e-99b9-871f5bfd5c79" containerID="434a1929d0adc58775e37e4c42628062d205cfb846f355095f66e49f55cf29ea" exitCode=0 Nov 25 17:00:38 crc kubenswrapper[4802]: I1125 17:00:38.985855 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khtpq" event={"ID":"015fc770-4272-466e-99b9-871f5bfd5c79","Type":"ContainerDied","Data":"434a1929d0adc58775e37e4c42628062d205cfb846f355095f66e49f55cf29ea"} Nov 25 17:00:38 crc kubenswrapper[4802]: I1125 17:00:38.986014 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khtpq" event={"ID":"015fc770-4272-466e-99b9-871f5bfd5c79","Type":"ContainerStarted","Data":"46c96fe55d7e196e913831a7cd76cdf4f10de2f95284a0b6b00a3167d0ae4e8b"} Nov 25 17:00:38 crc kubenswrapper[4802]: I1125 17:00:38.989537 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kcw4t" event={"ID":"97fc5986-61ea-49b8-a339-cd4526610431","Type":"ContainerStarted","Data":"01089cd39a46356073fc3a15d5c662cc214a932878fcd2fd6f91beecc6f1908e"} Nov 25 17:00:39 crc kubenswrapper[4802]: I1125 17:00:39.265010 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" Nov 25 17:00:39 crc kubenswrapper[4802]: I1125 17:00:39.393925 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d3f9606c-d757-40c8-a32e-52b5d9572f65-util\") pod \"d3f9606c-d757-40c8-a32e-52b5d9572f65\" (UID: \"d3f9606c-d757-40c8-a32e-52b5d9572f65\") " Nov 25 17:00:39 crc kubenswrapper[4802]: I1125 17:00:39.393980 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjrwv\" (UniqueName: \"kubernetes.io/projected/d3f9606c-d757-40c8-a32e-52b5d9572f65-kube-api-access-rjrwv\") pod \"d3f9606c-d757-40c8-a32e-52b5d9572f65\" (UID: \"d3f9606c-d757-40c8-a32e-52b5d9572f65\") " Nov 25 17:00:39 crc kubenswrapper[4802]: I1125 17:00:39.394015 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d3f9606c-d757-40c8-a32e-52b5d9572f65-bundle\") pod \"d3f9606c-d757-40c8-a32e-52b5d9572f65\" (UID: \"d3f9606c-d757-40c8-a32e-52b5d9572f65\") " Nov 25 17:00:39 crc kubenswrapper[4802]: I1125 17:00:39.394965 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3f9606c-d757-40c8-a32e-52b5d9572f65-bundle" (OuterVolumeSpecName: "bundle") pod "d3f9606c-d757-40c8-a32e-52b5d9572f65" (UID: "d3f9606c-d757-40c8-a32e-52b5d9572f65"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:00:39 crc kubenswrapper[4802]: I1125 17:00:39.400547 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3f9606c-d757-40c8-a32e-52b5d9572f65-kube-api-access-rjrwv" (OuterVolumeSpecName: "kube-api-access-rjrwv") pod "d3f9606c-d757-40c8-a32e-52b5d9572f65" (UID: "d3f9606c-d757-40c8-a32e-52b5d9572f65"). InnerVolumeSpecName "kube-api-access-rjrwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:00:39 crc kubenswrapper[4802]: I1125 17:00:39.410326 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3f9606c-d757-40c8-a32e-52b5d9572f65-util" (OuterVolumeSpecName: "util") pod "d3f9606c-d757-40c8-a32e-52b5d9572f65" (UID: "d3f9606c-d757-40c8-a32e-52b5d9572f65"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:00:39 crc kubenswrapper[4802]: I1125 17:00:39.495877 4802 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d3f9606c-d757-40c8-a32e-52b5d9572f65-util\") on node \"crc\" DevicePath \"\"" Nov 25 17:00:39 crc kubenswrapper[4802]: I1125 17:00:39.495926 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjrwv\" (UniqueName: \"kubernetes.io/projected/d3f9606c-d757-40c8-a32e-52b5d9572f65-kube-api-access-rjrwv\") on node \"crc\" DevicePath \"\"" Nov 25 17:00:39 crc kubenswrapper[4802]: I1125 17:00:39.495936 4802 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d3f9606c-d757-40c8-a32e-52b5d9572f65-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 17:00:39 crc kubenswrapper[4802]: I1125 17:00:39.996971 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" event={"ID":"d3f9606c-d757-40c8-a32e-52b5d9572f65","Type":"ContainerDied","Data":"7ba475c1dbd5e469700bddd9b20f7046eee733239d2cf4034c2dc7544dbd2b03"} Nov 25 17:00:39 crc kubenswrapper[4802]: I1125 17:00:39.997023 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ba475c1dbd5e469700bddd9b20f7046eee733239d2cf4034c2dc7544dbd2b03" Nov 25 17:00:39 crc kubenswrapper[4802]: I1125 17:00:39.996991 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc" Nov 25 17:00:39 crc kubenswrapper[4802]: I1125 17:00:39.998923 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khtpq" event={"ID":"015fc770-4272-466e-99b9-871f5bfd5c79","Type":"ContainerStarted","Data":"9ff6621531dc11bacde485696e3170868159e0d045483f79b3b2edf8dd211b37"} Nov 25 17:00:40 crc kubenswrapper[4802]: I1125 17:00:40.000616 4802 generic.go:334] "Generic (PLEG): container finished" podID="97fc5986-61ea-49b8-a339-cd4526610431" containerID="01089cd39a46356073fc3a15d5c662cc214a932878fcd2fd6f91beecc6f1908e" exitCode=0 Nov 25 17:00:40 crc kubenswrapper[4802]: I1125 17:00:40.000648 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kcw4t" event={"ID":"97fc5986-61ea-49b8-a339-cd4526610431","Type":"ContainerDied","Data":"01089cd39a46356073fc3a15d5c662cc214a932878fcd2fd6f91beecc6f1908e"} Nov 25 17:00:41 crc kubenswrapper[4802]: I1125 17:00:41.009058 4802 generic.go:334] "Generic (PLEG): container finished" podID="015fc770-4272-466e-99b9-871f5bfd5c79" containerID="9ff6621531dc11bacde485696e3170868159e0d045483f79b3b2edf8dd211b37" exitCode=0 Nov 25 17:00:41 crc kubenswrapper[4802]: I1125 17:00:41.009156 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khtpq" event={"ID":"015fc770-4272-466e-99b9-871f5bfd5c79","Type":"ContainerDied","Data":"9ff6621531dc11bacde485696e3170868159e0d045483f79b3b2edf8dd211b37"} Nov 25 17:00:41 crc kubenswrapper[4802]: I1125 17:00:41.012135 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kcw4t" event={"ID":"97fc5986-61ea-49b8-a339-cd4526610431","Type":"ContainerStarted","Data":"074ee674850879141220938f940d4cc3b0fa5f033180dd8ebed4ae5e323de462"} Nov 25 17:00:41 crc kubenswrapper[4802]: I1125 17:00:41.055429 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kcw4t" podStartSLOduration=2.571606267 podStartE2EDuration="5.05540456s" podCreationTimestamp="2025-11-25 17:00:36 +0000 UTC" firstStartedPulling="2025-11-25 17:00:37.974909236 +0000 UTC m=+821.119256422" lastFinishedPulling="2025-11-25 17:00:40.458707529 +0000 UTC m=+823.603054715" observedRunningTime="2025-11-25 17:00:41.053816816 +0000 UTC m=+824.198164002" watchObservedRunningTime="2025-11-25 17:00:41.05540456 +0000 UTC m=+824.199751746" Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.019407 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khtpq" event={"ID":"015fc770-4272-466e-99b9-871f5bfd5c79","Type":"ContainerStarted","Data":"bade429ae6ce71c679cfe27cae7f07e00cb0806c5a1af92ae373e9f381e8b3b9"} Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.038646 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-khtpq" podStartSLOduration=2.594600464 podStartE2EDuration="5.038629198s" podCreationTimestamp="2025-11-25 17:00:37 +0000 UTC" firstStartedPulling="2025-11-25 17:00:38.989365722 +0000 UTC m=+822.133712908" lastFinishedPulling="2025-11-25 17:00:41.433394456 +0000 UTC m=+824.577741642" observedRunningTime="2025-11-25 17:00:42.036620743 +0000 UTC m=+825.180967929" watchObservedRunningTime="2025-11-25 17:00:42.038629198 +0000 UTC m=+825.182976384" Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.098325 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x7mdl"] Nov 25 17:00:42 crc kubenswrapper[4802]: E1125 17:00:42.098650 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f9606c-d757-40c8-a32e-52b5d9572f65" containerName="pull" Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.098668 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f9606c-d757-40c8-a32e-52b5d9572f65" containerName="pull" Nov 25 17:00:42 crc kubenswrapper[4802]: E1125 17:00:42.098681 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f9606c-d757-40c8-a32e-52b5d9572f65" containerName="util" Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.098687 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f9606c-d757-40c8-a32e-52b5d9572f65" containerName="util" Nov 25 17:00:42 crc kubenswrapper[4802]: E1125 17:00:42.098700 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f9606c-d757-40c8-a32e-52b5d9572f65" containerName="extract" Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.098706 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f9606c-d757-40c8-a32e-52b5d9572f65" containerName="extract" Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.098842 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f9606c-d757-40c8-a32e-52b5d9572f65" containerName="extract" Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.099871 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7mdl" Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.110169 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7mdl"] Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.234404 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfszd\" (UniqueName: \"kubernetes.io/projected/955f36f6-4779-4e55-978d-177a8d1849b4-kube-api-access-dfszd\") pod \"redhat-operators-x7mdl\" (UID: \"955f36f6-4779-4e55-978d-177a8d1849b4\") " pod="openshift-marketplace/redhat-operators-x7mdl" Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.234475 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/955f36f6-4779-4e55-978d-177a8d1849b4-catalog-content\") pod \"redhat-operators-x7mdl\" (UID: \"955f36f6-4779-4e55-978d-177a8d1849b4\") " pod="openshift-marketplace/redhat-operators-x7mdl" Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.234505 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/955f36f6-4779-4e55-978d-177a8d1849b4-utilities\") pod \"redhat-operators-x7mdl\" (UID: \"955f36f6-4779-4e55-978d-177a8d1849b4\") " pod="openshift-marketplace/redhat-operators-x7mdl" Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.337328 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfszd\" (UniqueName: \"kubernetes.io/projected/955f36f6-4779-4e55-978d-177a8d1849b4-kube-api-access-dfszd\") pod \"redhat-operators-x7mdl\" (UID: \"955f36f6-4779-4e55-978d-177a8d1849b4\") " pod="openshift-marketplace/redhat-operators-x7mdl" Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.337643 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/955f36f6-4779-4e55-978d-177a8d1849b4-catalog-content\") pod \"redhat-operators-x7mdl\" (UID: \"955f36f6-4779-4e55-978d-177a8d1849b4\") " pod="openshift-marketplace/redhat-operators-x7mdl" Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.337740 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/955f36f6-4779-4e55-978d-177a8d1849b4-utilities\") pod \"redhat-operators-x7mdl\" (UID: \"955f36f6-4779-4e55-978d-177a8d1849b4\") " pod="openshift-marketplace/redhat-operators-x7mdl" Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.338245 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/955f36f6-4779-4e55-978d-177a8d1849b4-catalog-content\") pod \"redhat-operators-x7mdl\" (UID: \"955f36f6-4779-4e55-978d-177a8d1849b4\") " pod="openshift-marketplace/redhat-operators-x7mdl" Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.338368 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/955f36f6-4779-4e55-978d-177a8d1849b4-utilities\") pod \"redhat-operators-x7mdl\" (UID: \"955f36f6-4779-4e55-978d-177a8d1849b4\") " pod="openshift-marketplace/redhat-operators-x7mdl" Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.363257 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfszd\" (UniqueName: \"kubernetes.io/projected/955f36f6-4779-4e55-978d-177a8d1849b4-kube-api-access-dfszd\") pod \"redhat-operators-x7mdl\" (UID: \"955f36f6-4779-4e55-978d-177a8d1849b4\") " pod="openshift-marketplace/redhat-operators-x7mdl" Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.413761 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7mdl" Nov 25 17:00:42 crc kubenswrapper[4802]: I1125 17:00:42.833859 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7mdl"] Nov 25 17:00:43 crc kubenswrapper[4802]: I1125 17:00:43.026300 4802 generic.go:334] "Generic (PLEG): container finished" podID="955f36f6-4779-4e55-978d-177a8d1849b4" containerID="da5e8c5c8ee98472ff0df088821d3ebca4d55f1e915c73005509c641f7bbf636" exitCode=0 Nov 25 17:00:43 crc kubenswrapper[4802]: I1125 17:00:43.026352 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7mdl" event={"ID":"955f36f6-4779-4e55-978d-177a8d1849b4","Type":"ContainerDied","Data":"da5e8c5c8ee98472ff0df088821d3ebca4d55f1e915c73005509c641f7bbf636"} Nov 25 17:00:43 crc kubenswrapper[4802]: I1125 17:00:43.026650 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7mdl" event={"ID":"955f36f6-4779-4e55-978d-177a8d1849b4","Type":"ContainerStarted","Data":"871eab657b3fdd876620824f402f07b26c581f6243942a727c2c28ea122603ca"} Nov 25 17:00:44 crc kubenswrapper[4802]: I1125 17:00:44.038538 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7mdl" event={"ID":"955f36f6-4779-4e55-978d-177a8d1849b4","Type":"ContainerStarted","Data":"b7026a40692ff071051cd45f0911c3ae944f42db0b5f087cf5ff2dd525939fcc"} Nov 25 17:00:45 crc kubenswrapper[4802]: I1125 17:00:45.046664 4802 generic.go:334] "Generic (PLEG): container finished" podID="955f36f6-4779-4e55-978d-177a8d1849b4" containerID="b7026a40692ff071051cd45f0911c3ae944f42db0b5f087cf5ff2dd525939fcc" exitCode=0 Nov 25 17:00:45 crc kubenswrapper[4802]: I1125 17:00:45.046803 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7mdl" event={"ID":"955f36f6-4779-4e55-978d-177a8d1849b4","Type":"ContainerDied","Data":"b7026a40692ff071051cd45f0911c3ae944f42db0b5f087cf5ff2dd525939fcc"} Nov 25 17:00:45 crc kubenswrapper[4802]: I1125 17:00:45.524354 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc"] Nov 25 17:00:45 crc kubenswrapper[4802]: I1125 17:00:45.525334 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:00:45 crc kubenswrapper[4802]: I1125 17:00:45.527368 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 25 17:00:45 crc kubenswrapper[4802]: I1125 17:00:45.527640 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-mcsgb" Nov 25 17:00:45 crc kubenswrapper[4802]: I1125 17:00:45.528535 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Nov 25 17:00:45 crc kubenswrapper[4802]: I1125 17:00:45.549676 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc"] Nov 25 17:00:45 crc kubenswrapper[4802]: I1125 17:00:45.688607 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6p8s\" (UniqueName: \"kubernetes.io/projected/5004bc39-3a06-4842-9565-8501f35c8c4b-kube-api-access-s6p8s\") pod \"mariadb-operator-controller-manager-5959d99654-4j2qc\" (UID: \"5004bc39-3a06-4842-9565-8501f35c8c4b\") " pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:00:45 crc kubenswrapper[4802]: I1125 17:00:45.688690 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5004bc39-3a06-4842-9565-8501f35c8c4b-apiservice-cert\") pod \"mariadb-operator-controller-manager-5959d99654-4j2qc\" (UID: \"5004bc39-3a06-4842-9565-8501f35c8c4b\") " pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:00:45 crc kubenswrapper[4802]: I1125 17:00:45.688751 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5004bc39-3a06-4842-9565-8501f35c8c4b-webhook-cert\") pod \"mariadb-operator-controller-manager-5959d99654-4j2qc\" (UID: \"5004bc39-3a06-4842-9565-8501f35c8c4b\") " pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:00:45 crc kubenswrapper[4802]: I1125 17:00:45.791678 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6p8s\" (UniqueName: \"kubernetes.io/projected/5004bc39-3a06-4842-9565-8501f35c8c4b-kube-api-access-s6p8s\") pod \"mariadb-operator-controller-manager-5959d99654-4j2qc\" (UID: \"5004bc39-3a06-4842-9565-8501f35c8c4b\") " pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:00:45 crc kubenswrapper[4802]: I1125 17:00:45.791800 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5004bc39-3a06-4842-9565-8501f35c8c4b-apiservice-cert\") pod \"mariadb-operator-controller-manager-5959d99654-4j2qc\" (UID: \"5004bc39-3a06-4842-9565-8501f35c8c4b\") " pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:00:45 crc kubenswrapper[4802]: I1125 17:00:45.791831 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5004bc39-3a06-4842-9565-8501f35c8c4b-webhook-cert\") pod \"mariadb-operator-controller-manager-5959d99654-4j2qc\" (UID: \"5004bc39-3a06-4842-9565-8501f35c8c4b\") " pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:00:45 crc kubenswrapper[4802]: I1125 17:00:45.798879 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5004bc39-3a06-4842-9565-8501f35c8c4b-apiservice-cert\") pod \"mariadb-operator-controller-manager-5959d99654-4j2qc\" (UID: \"5004bc39-3a06-4842-9565-8501f35c8c4b\") " pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:00:45 crc kubenswrapper[4802]: I1125 17:00:45.799364 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5004bc39-3a06-4842-9565-8501f35c8c4b-webhook-cert\") pod \"mariadb-operator-controller-manager-5959d99654-4j2qc\" (UID: \"5004bc39-3a06-4842-9565-8501f35c8c4b\") " pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:00:45 crc kubenswrapper[4802]: I1125 17:00:45.811900 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6p8s\" (UniqueName: \"kubernetes.io/projected/5004bc39-3a06-4842-9565-8501f35c8c4b-kube-api-access-s6p8s\") pod \"mariadb-operator-controller-manager-5959d99654-4j2qc\" (UID: \"5004bc39-3a06-4842-9565-8501f35c8c4b\") " pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:00:45 crc kubenswrapper[4802]: I1125 17:00:45.845611 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:00:46 crc kubenswrapper[4802]: I1125 17:00:46.055410 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7mdl" event={"ID":"955f36f6-4779-4e55-978d-177a8d1849b4","Type":"ContainerStarted","Data":"3a1dbbb816e4b8af87726253eb24ec19fac9acf827bca2a75723c2ee8da4694c"} Nov 25 17:00:46 crc kubenswrapper[4802]: I1125 17:00:46.079615 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x7mdl" podStartSLOduration=1.352902718 podStartE2EDuration="4.079595561s" podCreationTimestamp="2025-11-25 17:00:42 +0000 UTC" firstStartedPulling="2025-11-25 17:00:43.028065944 +0000 UTC m=+826.172413130" lastFinishedPulling="2025-11-25 17:00:45.754758787 +0000 UTC m=+828.899105973" observedRunningTime="2025-11-25 17:00:46.078286465 +0000 UTC m=+829.222633661" watchObservedRunningTime="2025-11-25 17:00:46.079595561 +0000 UTC m=+829.223942737" Nov 25 17:00:46 crc kubenswrapper[4802]: I1125 17:00:46.259969 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc"] Nov 25 17:00:46 crc kubenswrapper[4802]: W1125 17:00:46.268104 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5004bc39_3a06_4842_9565_8501f35c8c4b.slice/crio-6964f4bfdf246de26c7b434d8202917d3112a0dde68052d97c5d6449a31fa7e1 WatchSource:0}: Error finding container 6964f4bfdf246de26c7b434d8202917d3112a0dde68052d97c5d6449a31fa7e1: Status 404 returned error can't find the container with id 6964f4bfdf246de26c7b434d8202917d3112a0dde68052d97c5d6449a31fa7e1 Nov 25 17:00:46 crc kubenswrapper[4802]: I1125 17:00:46.414530 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kcw4t" Nov 25 17:00:46 crc kubenswrapper[4802]: I1125 17:00:46.414577 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kcw4t" Nov 25 17:00:46 crc kubenswrapper[4802]: I1125 17:00:46.463329 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kcw4t" Nov 25 17:00:47 crc kubenswrapper[4802]: I1125 17:00:47.068923 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" event={"ID":"5004bc39-3a06-4842-9565-8501f35c8c4b","Type":"ContainerStarted","Data":"6964f4bfdf246de26c7b434d8202917d3112a0dde68052d97c5d6449a31fa7e1"} Nov 25 17:00:47 crc kubenswrapper[4802]: I1125 17:00:47.118289 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kcw4t" Nov 25 17:00:47 crc kubenswrapper[4802]: I1125 17:00:47.615729 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-khtpq" Nov 25 17:00:47 crc kubenswrapper[4802]: I1125 17:00:47.616055 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-khtpq" Nov 25 17:00:47 crc kubenswrapper[4802]: I1125 17:00:47.668422 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-khtpq" Nov 25 17:00:48 crc kubenswrapper[4802]: I1125 17:00:48.120418 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-khtpq" Nov 25 17:00:50 crc kubenswrapper[4802]: I1125 17:00:50.090451 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" event={"ID":"5004bc39-3a06-4842-9565-8501f35c8c4b","Type":"ContainerStarted","Data":"f36e88039dd0e0ed5ed95ac4f1f0d31d7927e4940038a72419458c2d43e546d4"} Nov 25 17:00:50 crc kubenswrapper[4802]: I1125 17:00:50.090819 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:00:50 crc kubenswrapper[4802]: I1125 17:00:50.107551 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" podStartSLOduration=1.865141795 podStartE2EDuration="5.107528051s" podCreationTimestamp="2025-11-25 17:00:45 +0000 UTC" firstStartedPulling="2025-11-25 17:00:46.270331106 +0000 UTC m=+829.414678292" lastFinishedPulling="2025-11-25 17:00:49.512717362 +0000 UTC m=+832.657064548" observedRunningTime="2025-11-25 17:00:50.106605036 +0000 UTC m=+833.250952232" watchObservedRunningTime="2025-11-25 17:00:50.107528051 +0000 UTC m=+833.251875237" Nov 25 17:00:51 crc kubenswrapper[4802]: I1125 17:00:51.687326 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kcw4t"] Nov 25 17:00:51 crc kubenswrapper[4802]: I1125 17:00:51.687598 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kcw4t" podUID="97fc5986-61ea-49b8-a339-cd4526610431" containerName="registry-server" containerID="cri-o://074ee674850879141220938f940d4cc3b0fa5f033180dd8ebed4ae5e323de462" gracePeriod=2 Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.059429 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kcw4t" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.100394 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-khtpq"] Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.100651 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-khtpq" podUID="015fc770-4272-466e-99b9-871f5bfd5c79" containerName="registry-server" containerID="cri-o://bade429ae6ce71c679cfe27cae7f07e00cb0806c5a1af92ae373e9f381e8b3b9" gracePeriod=2 Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.104093 4802 generic.go:334] "Generic (PLEG): container finished" podID="97fc5986-61ea-49b8-a339-cd4526610431" containerID="074ee674850879141220938f940d4cc3b0fa5f033180dd8ebed4ae5e323de462" exitCode=0 Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.104154 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kcw4t" event={"ID":"97fc5986-61ea-49b8-a339-cd4526610431","Type":"ContainerDied","Data":"074ee674850879141220938f940d4cc3b0fa5f033180dd8ebed4ae5e323de462"} Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.104179 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kcw4t" event={"ID":"97fc5986-61ea-49b8-a339-cd4526610431","Type":"ContainerDied","Data":"09ebfe05736cca36f6abdd1b3d81b71f64ff53bd80c02c4f02ce226f96a6779f"} Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.104196 4802 scope.go:117] "RemoveContainer" containerID="074ee674850879141220938f940d4cc3b0fa5f033180dd8ebed4ae5e323de462" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.104316 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kcw4t" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.121621 4802 scope.go:117] "RemoveContainer" containerID="01089cd39a46356073fc3a15d5c662cc214a932878fcd2fd6f91beecc6f1908e" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.137924 4802 scope.go:117] "RemoveContainer" containerID="154d8d9cd3d99be5e2246b3befcf5bb4460dd8a6e57aef574b90170c61ee48a1" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.184785 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97fc5986-61ea-49b8-a339-cd4526610431-utilities\") pod \"97fc5986-61ea-49b8-a339-cd4526610431\" (UID: \"97fc5986-61ea-49b8-a339-cd4526610431\") " Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.184948 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97fc5986-61ea-49b8-a339-cd4526610431-catalog-content\") pod \"97fc5986-61ea-49b8-a339-cd4526610431\" (UID: \"97fc5986-61ea-49b8-a339-cd4526610431\") " Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.185017 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9smdc\" (UniqueName: \"kubernetes.io/projected/97fc5986-61ea-49b8-a339-cd4526610431-kube-api-access-9smdc\") pod \"97fc5986-61ea-49b8-a339-cd4526610431\" (UID: \"97fc5986-61ea-49b8-a339-cd4526610431\") " Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.186384 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97fc5986-61ea-49b8-a339-cd4526610431-utilities" (OuterVolumeSpecName: "utilities") pod "97fc5986-61ea-49b8-a339-cd4526610431" (UID: "97fc5986-61ea-49b8-a339-cd4526610431"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.190468 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97fc5986-61ea-49b8-a339-cd4526610431-kube-api-access-9smdc" (OuterVolumeSpecName: "kube-api-access-9smdc") pod "97fc5986-61ea-49b8-a339-cd4526610431" (UID: "97fc5986-61ea-49b8-a339-cd4526610431"). InnerVolumeSpecName "kube-api-access-9smdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.201550 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97fc5986-61ea-49b8-a339-cd4526610431-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97fc5986-61ea-49b8-a339-cd4526610431" (UID: "97fc5986-61ea-49b8-a339-cd4526610431"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.253549 4802 scope.go:117] "RemoveContainer" containerID="074ee674850879141220938f940d4cc3b0fa5f033180dd8ebed4ae5e323de462" Nov 25 17:00:52 crc kubenswrapper[4802]: E1125 17:00:52.253975 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"074ee674850879141220938f940d4cc3b0fa5f033180dd8ebed4ae5e323de462\": container with ID starting with 074ee674850879141220938f940d4cc3b0fa5f033180dd8ebed4ae5e323de462 not found: ID does not exist" containerID="074ee674850879141220938f940d4cc3b0fa5f033180dd8ebed4ae5e323de462" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.254017 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"074ee674850879141220938f940d4cc3b0fa5f033180dd8ebed4ae5e323de462"} err="failed to get container status \"074ee674850879141220938f940d4cc3b0fa5f033180dd8ebed4ae5e323de462\": rpc error: code = NotFound desc = could not find container \"074ee674850879141220938f940d4cc3b0fa5f033180dd8ebed4ae5e323de462\": container with ID starting with 074ee674850879141220938f940d4cc3b0fa5f033180dd8ebed4ae5e323de462 not found: ID does not exist" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.254046 4802 scope.go:117] "RemoveContainer" containerID="01089cd39a46356073fc3a15d5c662cc214a932878fcd2fd6f91beecc6f1908e" Nov 25 17:00:52 crc kubenswrapper[4802]: E1125 17:00:52.254508 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01089cd39a46356073fc3a15d5c662cc214a932878fcd2fd6f91beecc6f1908e\": container with ID starting with 01089cd39a46356073fc3a15d5c662cc214a932878fcd2fd6f91beecc6f1908e not found: ID does not exist" containerID="01089cd39a46356073fc3a15d5c662cc214a932878fcd2fd6f91beecc6f1908e" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.254550 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01089cd39a46356073fc3a15d5c662cc214a932878fcd2fd6f91beecc6f1908e"} err="failed to get container status \"01089cd39a46356073fc3a15d5c662cc214a932878fcd2fd6f91beecc6f1908e\": rpc error: code = NotFound desc = could not find container \"01089cd39a46356073fc3a15d5c662cc214a932878fcd2fd6f91beecc6f1908e\": container with ID starting with 01089cd39a46356073fc3a15d5c662cc214a932878fcd2fd6f91beecc6f1908e not found: ID does not exist" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.254582 4802 scope.go:117] "RemoveContainer" containerID="154d8d9cd3d99be5e2246b3befcf5bb4460dd8a6e57aef574b90170c61ee48a1" Nov 25 17:00:52 crc kubenswrapper[4802]: E1125 17:00:52.254898 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"154d8d9cd3d99be5e2246b3befcf5bb4460dd8a6e57aef574b90170c61ee48a1\": container with ID starting with 154d8d9cd3d99be5e2246b3befcf5bb4460dd8a6e57aef574b90170c61ee48a1 not found: ID does not exist" containerID="154d8d9cd3d99be5e2246b3befcf5bb4460dd8a6e57aef574b90170c61ee48a1" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.254926 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"154d8d9cd3d99be5e2246b3befcf5bb4460dd8a6e57aef574b90170c61ee48a1"} err="failed to get container status \"154d8d9cd3d99be5e2246b3befcf5bb4460dd8a6e57aef574b90170c61ee48a1\": rpc error: code = NotFound desc = could not find container \"154d8d9cd3d99be5e2246b3befcf5bb4460dd8a6e57aef574b90170c61ee48a1\": container with ID starting with 154d8d9cd3d99be5e2246b3befcf5bb4460dd8a6e57aef574b90170c61ee48a1 not found: ID does not exist" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.286673 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97fc5986-61ea-49b8-a339-cd4526610431-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.286718 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9smdc\" (UniqueName: \"kubernetes.io/projected/97fc5986-61ea-49b8-a339-cd4526610431-kube-api-access-9smdc\") on node \"crc\" DevicePath \"\"" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.286733 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97fc5986-61ea-49b8-a339-cd4526610431-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.413988 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x7mdl" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.414027 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x7mdl" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.438041 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kcw4t"] Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.441016 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kcw4t"] Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.464947 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x7mdl" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.496435 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-khtpq" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.589830 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrpk9\" (UniqueName: \"kubernetes.io/projected/015fc770-4272-466e-99b9-871f5bfd5c79-kube-api-access-zrpk9\") pod \"015fc770-4272-466e-99b9-871f5bfd5c79\" (UID: \"015fc770-4272-466e-99b9-871f5bfd5c79\") " Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.589879 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/015fc770-4272-466e-99b9-871f5bfd5c79-utilities\") pod \"015fc770-4272-466e-99b9-871f5bfd5c79\" (UID: \"015fc770-4272-466e-99b9-871f5bfd5c79\") " Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.589936 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/015fc770-4272-466e-99b9-871f5bfd5c79-catalog-content\") pod \"015fc770-4272-466e-99b9-871f5bfd5c79\" (UID: \"015fc770-4272-466e-99b9-871f5bfd5c79\") " Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.590811 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/015fc770-4272-466e-99b9-871f5bfd5c79-utilities" (OuterVolumeSpecName: "utilities") pod "015fc770-4272-466e-99b9-871f5bfd5c79" (UID: "015fc770-4272-466e-99b9-871f5bfd5c79"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.594380 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/015fc770-4272-466e-99b9-871f5bfd5c79-kube-api-access-zrpk9" (OuterVolumeSpecName: "kube-api-access-zrpk9") pod "015fc770-4272-466e-99b9-871f5bfd5c79" (UID: "015fc770-4272-466e-99b9-871f5bfd5c79"). InnerVolumeSpecName "kube-api-access-zrpk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.636924 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/015fc770-4272-466e-99b9-871f5bfd5c79-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "015fc770-4272-466e-99b9-871f5bfd5c79" (UID: "015fc770-4272-466e-99b9-871f5bfd5c79"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.691701 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/015fc770-4272-466e-99b9-871f5bfd5c79-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.691764 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/015fc770-4272-466e-99b9-871f5bfd5c79-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 17:00:52 crc kubenswrapper[4802]: I1125 17:00:52.691778 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrpk9\" (UniqueName: \"kubernetes.io/projected/015fc770-4272-466e-99b9-871f5bfd5c79-kube-api-access-zrpk9\") on node \"crc\" DevicePath \"\"" Nov 25 17:00:53 crc kubenswrapper[4802]: I1125 17:00:53.114046 4802 generic.go:334] "Generic (PLEG): container finished" podID="015fc770-4272-466e-99b9-871f5bfd5c79" containerID="bade429ae6ce71c679cfe27cae7f07e00cb0806c5a1af92ae373e9f381e8b3b9" exitCode=0 Nov 25 17:00:53 crc kubenswrapper[4802]: I1125 17:00:53.114148 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-khtpq" Nov 25 17:00:53 crc kubenswrapper[4802]: I1125 17:00:53.114170 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khtpq" event={"ID":"015fc770-4272-466e-99b9-871f5bfd5c79","Type":"ContainerDied","Data":"bade429ae6ce71c679cfe27cae7f07e00cb0806c5a1af92ae373e9f381e8b3b9"} Nov 25 17:00:53 crc kubenswrapper[4802]: I1125 17:00:53.114918 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khtpq" event={"ID":"015fc770-4272-466e-99b9-871f5bfd5c79","Type":"ContainerDied","Data":"46c96fe55d7e196e913831a7cd76cdf4f10de2f95284a0b6b00a3167d0ae4e8b"} Nov 25 17:00:53 crc kubenswrapper[4802]: I1125 17:00:53.114952 4802 scope.go:117] "RemoveContainer" containerID="bade429ae6ce71c679cfe27cae7f07e00cb0806c5a1af92ae373e9f381e8b3b9" Nov 25 17:00:53 crc kubenswrapper[4802]: I1125 17:00:53.136856 4802 scope.go:117] "RemoveContainer" containerID="9ff6621531dc11bacde485696e3170868159e0d045483f79b3b2edf8dd211b37" Nov 25 17:00:53 crc kubenswrapper[4802]: I1125 17:00:53.148426 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-khtpq"] Nov 25 17:00:53 crc kubenswrapper[4802]: I1125 17:00:53.153339 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-khtpq"] Nov 25 17:00:53 crc kubenswrapper[4802]: I1125 17:00:53.167252 4802 scope.go:117] "RemoveContainer" containerID="434a1929d0adc58775e37e4c42628062d205cfb846f355095f66e49f55cf29ea" Nov 25 17:00:53 crc kubenswrapper[4802]: I1125 17:00:53.170784 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x7mdl" Nov 25 17:00:53 crc kubenswrapper[4802]: I1125 17:00:53.191782 4802 scope.go:117] "RemoveContainer" containerID="bade429ae6ce71c679cfe27cae7f07e00cb0806c5a1af92ae373e9f381e8b3b9" Nov 25 17:00:53 crc kubenswrapper[4802]: E1125 17:00:53.192268 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bade429ae6ce71c679cfe27cae7f07e00cb0806c5a1af92ae373e9f381e8b3b9\": container with ID starting with bade429ae6ce71c679cfe27cae7f07e00cb0806c5a1af92ae373e9f381e8b3b9 not found: ID does not exist" containerID="bade429ae6ce71c679cfe27cae7f07e00cb0806c5a1af92ae373e9f381e8b3b9" Nov 25 17:00:53 crc kubenswrapper[4802]: I1125 17:00:53.192300 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bade429ae6ce71c679cfe27cae7f07e00cb0806c5a1af92ae373e9f381e8b3b9"} err="failed to get container status \"bade429ae6ce71c679cfe27cae7f07e00cb0806c5a1af92ae373e9f381e8b3b9\": rpc error: code = NotFound desc = could not find container \"bade429ae6ce71c679cfe27cae7f07e00cb0806c5a1af92ae373e9f381e8b3b9\": container with ID starting with bade429ae6ce71c679cfe27cae7f07e00cb0806c5a1af92ae373e9f381e8b3b9 not found: ID does not exist" Nov 25 17:00:53 crc kubenswrapper[4802]: I1125 17:00:53.192321 4802 scope.go:117] "RemoveContainer" containerID="9ff6621531dc11bacde485696e3170868159e0d045483f79b3b2edf8dd211b37" Nov 25 17:00:53 crc kubenswrapper[4802]: E1125 17:00:53.192804 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ff6621531dc11bacde485696e3170868159e0d045483f79b3b2edf8dd211b37\": container with ID starting with 9ff6621531dc11bacde485696e3170868159e0d045483f79b3b2edf8dd211b37 not found: ID does not exist" containerID="9ff6621531dc11bacde485696e3170868159e0d045483f79b3b2edf8dd211b37" Nov 25 17:00:53 crc kubenswrapper[4802]: I1125 17:00:53.192835 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ff6621531dc11bacde485696e3170868159e0d045483f79b3b2edf8dd211b37"} err="failed to get container status \"9ff6621531dc11bacde485696e3170868159e0d045483f79b3b2edf8dd211b37\": rpc error: code = NotFound desc = could not find container \"9ff6621531dc11bacde485696e3170868159e0d045483f79b3b2edf8dd211b37\": container with ID starting with 9ff6621531dc11bacde485696e3170868159e0d045483f79b3b2edf8dd211b37 not found: ID does not exist" Nov 25 17:00:53 crc kubenswrapper[4802]: I1125 17:00:53.192856 4802 scope.go:117] "RemoveContainer" containerID="434a1929d0adc58775e37e4c42628062d205cfb846f355095f66e49f55cf29ea" Nov 25 17:00:53 crc kubenswrapper[4802]: E1125 17:00:53.193513 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"434a1929d0adc58775e37e4c42628062d205cfb846f355095f66e49f55cf29ea\": container with ID starting with 434a1929d0adc58775e37e4c42628062d205cfb846f355095f66e49f55cf29ea not found: ID does not exist" containerID="434a1929d0adc58775e37e4c42628062d205cfb846f355095f66e49f55cf29ea" Nov 25 17:00:53 crc kubenswrapper[4802]: I1125 17:00:53.193533 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"434a1929d0adc58775e37e4c42628062d205cfb846f355095f66e49f55cf29ea"} err="failed to get container status \"434a1929d0adc58775e37e4c42628062d205cfb846f355095f66e49f55cf29ea\": rpc error: code = NotFound desc = could not find container \"434a1929d0adc58775e37e4c42628062d205cfb846f355095f66e49f55cf29ea\": container with ID starting with 434a1929d0adc58775e37e4c42628062d205cfb846f355095f66e49f55cf29ea not found: ID does not exist" Nov 25 17:00:53 crc kubenswrapper[4802]: I1125 17:00:53.513050 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="015fc770-4272-466e-99b9-871f5bfd5c79" path="/var/lib/kubelet/pods/015fc770-4272-466e-99b9-871f5bfd5c79/volumes" Nov 25 17:00:53 crc kubenswrapper[4802]: I1125 17:00:53.513926 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97fc5986-61ea-49b8-a339-cd4526610431" path="/var/lib/kubelet/pods/97fc5986-61ea-49b8-a339-cd4526610431/volumes" Nov 25 17:00:55 crc kubenswrapper[4802]: I1125 17:00:55.286446 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x7mdl"] Nov 25 17:00:55 crc kubenswrapper[4802]: I1125 17:00:55.287781 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x7mdl" podUID="955f36f6-4779-4e55-978d-177a8d1849b4" containerName="registry-server" containerID="cri-o://3a1dbbb816e4b8af87726253eb24ec19fac9acf827bca2a75723c2ee8da4694c" gracePeriod=2 Nov 25 17:00:55 crc kubenswrapper[4802]: I1125 17:00:55.864346 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:00:58 crc kubenswrapper[4802]: I1125 17:00:58.149699 4802 generic.go:334] "Generic (PLEG): container finished" podID="955f36f6-4779-4e55-978d-177a8d1849b4" containerID="3a1dbbb816e4b8af87726253eb24ec19fac9acf827bca2a75723c2ee8da4694c" exitCode=0 Nov 25 17:00:58 crc kubenswrapper[4802]: I1125 17:00:58.149789 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7mdl" event={"ID":"955f36f6-4779-4e55-978d-177a8d1849b4","Type":"ContainerDied","Data":"3a1dbbb816e4b8af87726253eb24ec19fac9acf827bca2a75723c2ee8da4694c"} Nov 25 17:00:58 crc kubenswrapper[4802]: I1125 17:00:58.526119 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7mdl" Nov 25 17:00:58 crc kubenswrapper[4802]: I1125 17:00:58.666336 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/955f36f6-4779-4e55-978d-177a8d1849b4-utilities\") pod \"955f36f6-4779-4e55-978d-177a8d1849b4\" (UID: \"955f36f6-4779-4e55-978d-177a8d1849b4\") " Nov 25 17:00:58 crc kubenswrapper[4802]: I1125 17:00:58.666429 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/955f36f6-4779-4e55-978d-177a8d1849b4-catalog-content\") pod \"955f36f6-4779-4e55-978d-177a8d1849b4\" (UID: \"955f36f6-4779-4e55-978d-177a8d1849b4\") " Nov 25 17:00:58 crc kubenswrapper[4802]: I1125 17:00:58.666467 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfszd\" (UniqueName: \"kubernetes.io/projected/955f36f6-4779-4e55-978d-177a8d1849b4-kube-api-access-dfszd\") pod \"955f36f6-4779-4e55-978d-177a8d1849b4\" (UID: \"955f36f6-4779-4e55-978d-177a8d1849b4\") " Nov 25 17:00:58 crc kubenswrapper[4802]: I1125 17:00:58.667464 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/955f36f6-4779-4e55-978d-177a8d1849b4-utilities" (OuterVolumeSpecName: "utilities") pod "955f36f6-4779-4e55-978d-177a8d1849b4" (UID: "955f36f6-4779-4e55-978d-177a8d1849b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:00:58 crc kubenswrapper[4802]: I1125 17:00:58.672734 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/955f36f6-4779-4e55-978d-177a8d1849b4-kube-api-access-dfszd" (OuterVolumeSpecName: "kube-api-access-dfszd") pod "955f36f6-4779-4e55-978d-177a8d1849b4" (UID: "955f36f6-4779-4e55-978d-177a8d1849b4"). InnerVolumeSpecName "kube-api-access-dfszd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:00:58 crc kubenswrapper[4802]: I1125 17:00:58.753393 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/955f36f6-4779-4e55-978d-177a8d1849b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "955f36f6-4779-4e55-978d-177a8d1849b4" (UID: "955f36f6-4779-4e55-978d-177a8d1849b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:00:58 crc kubenswrapper[4802]: I1125 17:00:58.767698 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/955f36f6-4779-4e55-978d-177a8d1849b4-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 17:00:58 crc kubenswrapper[4802]: I1125 17:00:58.767754 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/955f36f6-4779-4e55-978d-177a8d1849b4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 17:00:58 crc kubenswrapper[4802]: I1125 17:00:58.767765 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfszd\" (UniqueName: \"kubernetes.io/projected/955f36f6-4779-4e55-978d-177a8d1849b4-kube-api-access-dfszd\") on node \"crc\" DevicePath \"\"" Nov 25 17:00:59 crc kubenswrapper[4802]: I1125 17:00:59.158635 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7mdl" event={"ID":"955f36f6-4779-4e55-978d-177a8d1849b4","Type":"ContainerDied","Data":"871eab657b3fdd876620824f402f07b26c581f6243942a727c2c28ea122603ca"} Nov 25 17:00:59 crc kubenswrapper[4802]: I1125 17:00:59.158708 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7mdl" Nov 25 17:00:59 crc kubenswrapper[4802]: I1125 17:00:59.159436 4802 scope.go:117] "RemoveContainer" containerID="3a1dbbb816e4b8af87726253eb24ec19fac9acf827bca2a75723c2ee8da4694c" Nov 25 17:00:59 crc kubenswrapper[4802]: I1125 17:00:59.187845 4802 scope.go:117] "RemoveContainer" containerID="b7026a40692ff071051cd45f0911c3ae944f42db0b5f087cf5ff2dd525939fcc" Nov 25 17:00:59 crc kubenswrapper[4802]: I1125 17:00:59.190211 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x7mdl"] Nov 25 17:00:59 crc kubenswrapper[4802]: I1125 17:00:59.193951 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x7mdl"] Nov 25 17:00:59 crc kubenswrapper[4802]: I1125 17:00:59.206763 4802 scope.go:117] "RemoveContainer" containerID="da5e8c5c8ee98472ff0df088821d3ebca4d55f1e915c73005509c641f7bbf636" Nov 25 17:00:59 crc kubenswrapper[4802]: I1125 17:00:59.513103 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="955f36f6-4779-4e55-978d-177a8d1849b4" path="/var/lib/kubelet/pods/955f36f6-4779-4e55-978d-177a8d1849b4/volumes" Nov 25 17:01:01 crc kubenswrapper[4802]: I1125 17:01:01.894273 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-khg9j"] Nov 25 17:01:01 crc kubenswrapper[4802]: E1125 17:01:01.894803 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97fc5986-61ea-49b8-a339-cd4526610431" containerName="extract-content" Nov 25 17:01:01 crc kubenswrapper[4802]: I1125 17:01:01.894819 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="97fc5986-61ea-49b8-a339-cd4526610431" containerName="extract-content" Nov 25 17:01:01 crc kubenswrapper[4802]: E1125 17:01:01.894831 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="015fc770-4272-466e-99b9-871f5bfd5c79" containerName="extract-content" Nov 25 17:01:01 crc kubenswrapper[4802]: I1125 17:01:01.894837 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="015fc770-4272-466e-99b9-871f5bfd5c79" containerName="extract-content" Nov 25 17:01:01 crc kubenswrapper[4802]: E1125 17:01:01.894852 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97fc5986-61ea-49b8-a339-cd4526610431" containerName="extract-utilities" Nov 25 17:01:01 crc kubenswrapper[4802]: I1125 17:01:01.894857 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="97fc5986-61ea-49b8-a339-cd4526610431" containerName="extract-utilities" Nov 25 17:01:01 crc kubenswrapper[4802]: E1125 17:01:01.894866 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="955f36f6-4779-4e55-978d-177a8d1849b4" containerName="extract-content" Nov 25 17:01:01 crc kubenswrapper[4802]: I1125 17:01:01.894873 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="955f36f6-4779-4e55-978d-177a8d1849b4" containerName="extract-content" Nov 25 17:01:01 crc kubenswrapper[4802]: E1125 17:01:01.894882 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="015fc770-4272-466e-99b9-871f5bfd5c79" containerName="registry-server" Nov 25 17:01:01 crc kubenswrapper[4802]: I1125 17:01:01.894888 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="015fc770-4272-466e-99b9-871f5bfd5c79" containerName="registry-server" Nov 25 17:01:01 crc kubenswrapper[4802]: E1125 17:01:01.894896 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="955f36f6-4779-4e55-978d-177a8d1849b4" containerName="extract-utilities" Nov 25 17:01:01 crc kubenswrapper[4802]: I1125 17:01:01.894903 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="955f36f6-4779-4e55-978d-177a8d1849b4" containerName="extract-utilities" Nov 25 17:01:01 crc kubenswrapper[4802]: E1125 17:01:01.894910 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="955f36f6-4779-4e55-978d-177a8d1849b4" containerName="registry-server" Nov 25 17:01:01 crc kubenswrapper[4802]: I1125 17:01:01.894915 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="955f36f6-4779-4e55-978d-177a8d1849b4" containerName="registry-server" Nov 25 17:01:01 crc kubenswrapper[4802]: E1125 17:01:01.894927 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97fc5986-61ea-49b8-a339-cd4526610431" containerName="registry-server" Nov 25 17:01:01 crc kubenswrapper[4802]: I1125 17:01:01.894933 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="97fc5986-61ea-49b8-a339-cd4526610431" containerName="registry-server" Nov 25 17:01:01 crc kubenswrapper[4802]: E1125 17:01:01.894942 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="015fc770-4272-466e-99b9-871f5bfd5c79" containerName="extract-utilities" Nov 25 17:01:01 crc kubenswrapper[4802]: I1125 17:01:01.894947 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="015fc770-4272-466e-99b9-871f5bfd5c79" containerName="extract-utilities" Nov 25 17:01:01 crc kubenswrapper[4802]: I1125 17:01:01.895062 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="955f36f6-4779-4e55-978d-177a8d1849b4" containerName="registry-server" Nov 25 17:01:01 crc kubenswrapper[4802]: I1125 17:01:01.895074 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="97fc5986-61ea-49b8-a339-cd4526610431" containerName="registry-server" Nov 25 17:01:01 crc kubenswrapper[4802]: I1125 17:01:01.895084 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="015fc770-4272-466e-99b9-871f5bfd5c79" containerName="registry-server" Nov 25 17:01:01 crc kubenswrapper[4802]: I1125 17:01:01.895523 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-khg9j" Nov 25 17:01:01 crc kubenswrapper[4802]: I1125 17:01:01.898096 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-rxfjg" Nov 25 17:01:01 crc kubenswrapper[4802]: I1125 17:01:01.907980 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-khg9j"] Nov 25 17:01:02 crc kubenswrapper[4802]: I1125 17:01:02.012019 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk99g\" (UniqueName: \"kubernetes.io/projected/f8f50f91-e0de-42a8-b9b2-4b4eff04bd74-kube-api-access-lk99g\") pod \"infra-operator-index-khg9j\" (UID: \"f8f50f91-e0de-42a8-b9b2-4b4eff04bd74\") " pod="openstack-operators/infra-operator-index-khg9j" Nov 25 17:01:02 crc kubenswrapper[4802]: I1125 17:01:02.113694 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk99g\" (UniqueName: \"kubernetes.io/projected/f8f50f91-e0de-42a8-b9b2-4b4eff04bd74-kube-api-access-lk99g\") pod \"infra-operator-index-khg9j\" (UID: \"f8f50f91-e0de-42a8-b9b2-4b4eff04bd74\") " pod="openstack-operators/infra-operator-index-khg9j" Nov 25 17:01:02 crc kubenswrapper[4802]: I1125 17:01:02.138050 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk99g\" (UniqueName: \"kubernetes.io/projected/f8f50f91-e0de-42a8-b9b2-4b4eff04bd74-kube-api-access-lk99g\") pod \"infra-operator-index-khg9j\" (UID: \"f8f50f91-e0de-42a8-b9b2-4b4eff04bd74\") " pod="openstack-operators/infra-operator-index-khg9j" Nov 25 17:01:02 crc kubenswrapper[4802]: I1125 17:01:02.211227 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-khg9j" Nov 25 17:01:02 crc kubenswrapper[4802]: I1125 17:01:02.620071 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-khg9j"] Nov 25 17:01:03 crc kubenswrapper[4802]: I1125 17:01:03.190287 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-khg9j" event={"ID":"f8f50f91-e0de-42a8-b9b2-4b4eff04bd74","Type":"ContainerStarted","Data":"39f23f5070bcee575bb73fce5649de3cc300d5d562d1600dc2016951deafd5ce"} Nov 25 17:01:04 crc kubenswrapper[4802]: I1125 17:01:04.207077 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-khg9j" event={"ID":"f8f50f91-e0de-42a8-b9b2-4b4eff04bd74","Type":"ContainerStarted","Data":"aca50939279e35b47b57eb496b347507e410a0bc9e28196bcc8d912473bf5528"} Nov 25 17:01:04 crc kubenswrapper[4802]: I1125 17:01:04.228257 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-khg9j" podStartSLOduration=2.094734713 podStartE2EDuration="3.228230752s" podCreationTimestamp="2025-11-25 17:01:01 +0000 UTC" firstStartedPulling="2025-11-25 17:01:02.627734042 +0000 UTC m=+845.772081228" lastFinishedPulling="2025-11-25 17:01:03.761230081 +0000 UTC m=+846.905577267" observedRunningTime="2025-11-25 17:01:04.220837013 +0000 UTC m=+847.365184199" watchObservedRunningTime="2025-11-25 17:01:04.228230752 +0000 UTC m=+847.372577958" Nov 25 17:01:12 crc kubenswrapper[4802]: I1125 17:01:12.212220 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-khg9j" Nov 25 17:01:12 crc kubenswrapper[4802]: I1125 17:01:12.212795 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-khg9j" Nov 25 17:01:12 crc kubenswrapper[4802]: I1125 17:01:12.237386 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-khg9j" Nov 25 17:01:12 crc kubenswrapper[4802]: I1125 17:01:12.279023 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-khg9j" Nov 25 17:01:15 crc kubenswrapper[4802]: I1125 17:01:15.128667 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn"] Nov 25 17:01:15 crc kubenswrapper[4802]: I1125 17:01:15.129845 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" Nov 25 17:01:15 crc kubenswrapper[4802]: I1125 17:01:15.131915 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-t7662" Nov 25 17:01:15 crc kubenswrapper[4802]: I1125 17:01:15.139467 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn"] Nov 25 17:01:15 crc kubenswrapper[4802]: I1125 17:01:15.288370 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbq59\" (UniqueName: \"kubernetes.io/projected/f6472f81-bf27-4839-a0a4-d7baa206c138-kube-api-access-bbq59\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn\" (UID: \"f6472f81-bf27-4839-a0a4-d7baa206c138\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" Nov 25 17:01:15 crc kubenswrapper[4802]: I1125 17:01:15.288635 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f6472f81-bf27-4839-a0a4-d7baa206c138-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn\" (UID: \"f6472f81-bf27-4839-a0a4-d7baa206c138\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" Nov 25 17:01:15 crc kubenswrapper[4802]: I1125 17:01:15.288689 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f6472f81-bf27-4839-a0a4-d7baa206c138-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn\" (UID: \"f6472f81-bf27-4839-a0a4-d7baa206c138\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" Nov 25 17:01:15 crc kubenswrapper[4802]: I1125 17:01:15.390230 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbq59\" (UniqueName: \"kubernetes.io/projected/f6472f81-bf27-4839-a0a4-d7baa206c138-kube-api-access-bbq59\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn\" (UID: \"f6472f81-bf27-4839-a0a4-d7baa206c138\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" Nov 25 17:01:15 crc kubenswrapper[4802]: I1125 17:01:15.390321 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f6472f81-bf27-4839-a0a4-d7baa206c138-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn\" (UID: \"f6472f81-bf27-4839-a0a4-d7baa206c138\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" Nov 25 17:01:15 crc kubenswrapper[4802]: I1125 17:01:15.390396 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f6472f81-bf27-4839-a0a4-d7baa206c138-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn\" (UID: \"f6472f81-bf27-4839-a0a4-d7baa206c138\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" Nov 25 17:01:15 crc kubenswrapper[4802]: I1125 17:01:15.390997 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f6472f81-bf27-4839-a0a4-d7baa206c138-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn\" (UID: \"f6472f81-bf27-4839-a0a4-d7baa206c138\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" Nov 25 17:01:15 crc kubenswrapper[4802]: I1125 17:01:15.391295 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f6472f81-bf27-4839-a0a4-d7baa206c138-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn\" (UID: \"f6472f81-bf27-4839-a0a4-d7baa206c138\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" Nov 25 17:01:15 crc kubenswrapper[4802]: I1125 17:01:15.410348 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbq59\" (UniqueName: \"kubernetes.io/projected/f6472f81-bf27-4839-a0a4-d7baa206c138-kube-api-access-bbq59\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn\" (UID: \"f6472f81-bf27-4839-a0a4-d7baa206c138\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" Nov 25 17:01:15 crc kubenswrapper[4802]: I1125 17:01:15.456110 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" Nov 25 17:01:15 crc kubenswrapper[4802]: I1125 17:01:15.853556 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn"] Nov 25 17:01:15 crc kubenswrapper[4802]: W1125 17:01:15.861050 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6472f81_bf27_4839_a0a4_d7baa206c138.slice/crio-8f51687affcf774212f5f0f640382d9c4f42fb6f95dd7a16e349262abdb131fc WatchSource:0}: Error finding container 8f51687affcf774212f5f0f640382d9c4f42fb6f95dd7a16e349262abdb131fc: Status 404 returned error can't find the container with id 8f51687affcf774212f5f0f640382d9c4f42fb6f95dd7a16e349262abdb131fc Nov 25 17:01:16 crc kubenswrapper[4802]: I1125 17:01:16.274705 4802 generic.go:334] "Generic (PLEG): container finished" podID="f6472f81-bf27-4839-a0a4-d7baa206c138" containerID="171741460a7c246886b234e9f5254d1fe098231abc9fdfc8c33c3cc67ace8715" exitCode=0 Nov 25 17:01:16 crc kubenswrapper[4802]: I1125 17:01:16.274748 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" event={"ID":"f6472f81-bf27-4839-a0a4-d7baa206c138","Type":"ContainerDied","Data":"171741460a7c246886b234e9f5254d1fe098231abc9fdfc8c33c3cc67ace8715"} Nov 25 17:01:16 crc kubenswrapper[4802]: I1125 17:01:16.274810 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" event={"ID":"f6472f81-bf27-4839-a0a4-d7baa206c138","Type":"ContainerStarted","Data":"8f51687affcf774212f5f0f640382d9c4f42fb6f95dd7a16e349262abdb131fc"} Nov 25 17:01:17 crc kubenswrapper[4802]: I1125 17:01:17.282318 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" event={"ID":"f6472f81-bf27-4839-a0a4-d7baa206c138","Type":"ContainerStarted","Data":"3695819b4103bf0336d50eff684c058b61615245c551530d0ef312a2f9b99435"} Nov 25 17:01:18 crc kubenswrapper[4802]: I1125 17:01:18.289681 4802 generic.go:334] "Generic (PLEG): container finished" podID="f6472f81-bf27-4839-a0a4-d7baa206c138" containerID="3695819b4103bf0336d50eff684c058b61615245c551530d0ef312a2f9b99435" exitCode=0 Nov 25 17:01:18 crc kubenswrapper[4802]: I1125 17:01:18.289764 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" event={"ID":"f6472f81-bf27-4839-a0a4-d7baa206c138","Type":"ContainerDied","Data":"3695819b4103bf0336d50eff684c058b61615245c551530d0ef312a2f9b99435"} Nov 25 17:01:19 crc kubenswrapper[4802]: I1125 17:01:19.300275 4802 generic.go:334] "Generic (PLEG): container finished" podID="f6472f81-bf27-4839-a0a4-d7baa206c138" containerID="6584c754107f49ef5affc31210a8707a8293029fd1b252a25677e9673a67b287" exitCode=0 Nov 25 17:01:19 crc kubenswrapper[4802]: I1125 17:01:19.300371 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" event={"ID":"f6472f81-bf27-4839-a0a4-d7baa206c138","Type":"ContainerDied","Data":"6584c754107f49ef5affc31210a8707a8293029fd1b252a25677e9673a67b287"} Nov 25 17:01:20 crc kubenswrapper[4802]: I1125 17:01:20.800944 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" Nov 25 17:01:20 crc kubenswrapper[4802]: I1125 17:01:20.866808 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbq59\" (UniqueName: \"kubernetes.io/projected/f6472f81-bf27-4839-a0a4-d7baa206c138-kube-api-access-bbq59\") pod \"f6472f81-bf27-4839-a0a4-d7baa206c138\" (UID: \"f6472f81-bf27-4839-a0a4-d7baa206c138\") " Nov 25 17:01:20 crc kubenswrapper[4802]: I1125 17:01:20.866849 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f6472f81-bf27-4839-a0a4-d7baa206c138-util\") pod \"f6472f81-bf27-4839-a0a4-d7baa206c138\" (UID: \"f6472f81-bf27-4839-a0a4-d7baa206c138\") " Nov 25 17:01:20 crc kubenswrapper[4802]: I1125 17:01:20.866874 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f6472f81-bf27-4839-a0a4-d7baa206c138-bundle\") pod \"f6472f81-bf27-4839-a0a4-d7baa206c138\" (UID: \"f6472f81-bf27-4839-a0a4-d7baa206c138\") " Nov 25 17:01:20 crc kubenswrapper[4802]: I1125 17:01:20.868048 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6472f81-bf27-4839-a0a4-d7baa206c138-bundle" (OuterVolumeSpecName: "bundle") pod "f6472f81-bf27-4839-a0a4-d7baa206c138" (UID: "f6472f81-bf27-4839-a0a4-d7baa206c138"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:01:20 crc kubenswrapper[4802]: I1125 17:01:20.872627 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6472f81-bf27-4839-a0a4-d7baa206c138-kube-api-access-bbq59" (OuterVolumeSpecName: "kube-api-access-bbq59") pod "f6472f81-bf27-4839-a0a4-d7baa206c138" (UID: "f6472f81-bf27-4839-a0a4-d7baa206c138"). InnerVolumeSpecName "kube-api-access-bbq59". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:01:20 crc kubenswrapper[4802]: I1125 17:01:20.881960 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6472f81-bf27-4839-a0a4-d7baa206c138-util" (OuterVolumeSpecName: "util") pod "f6472f81-bf27-4839-a0a4-d7baa206c138" (UID: "f6472f81-bf27-4839-a0a4-d7baa206c138"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:01:20 crc kubenswrapper[4802]: I1125 17:01:20.968469 4802 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f6472f81-bf27-4839-a0a4-d7baa206c138-util\") on node \"crc\" DevicePath \"\"" Nov 25 17:01:20 crc kubenswrapper[4802]: I1125 17:01:20.968502 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbq59\" (UniqueName: \"kubernetes.io/projected/f6472f81-bf27-4839-a0a4-d7baa206c138-kube-api-access-bbq59\") on node \"crc\" DevicePath \"\"" Nov 25 17:01:20 crc kubenswrapper[4802]: I1125 17:01:20.968516 4802 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f6472f81-bf27-4839-a0a4-d7baa206c138-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 17:01:21 crc kubenswrapper[4802]: I1125 17:01:21.317390 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" event={"ID":"f6472f81-bf27-4839-a0a4-d7baa206c138","Type":"ContainerDied","Data":"8f51687affcf774212f5f0f640382d9c4f42fb6f95dd7a16e349262abdb131fc"} Nov 25 17:01:21 crc kubenswrapper[4802]: I1125 17:01:21.317449 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f51687affcf774212f5f0f640382d9c4f42fb6f95dd7a16e349262abdb131fc" Nov 25 17:01:21 crc kubenswrapper[4802]: I1125 17:01:21.317446 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.087431 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf"] Nov 25 17:01:28 crc kubenswrapper[4802]: E1125 17:01:28.088117 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6472f81-bf27-4839-a0a4-d7baa206c138" containerName="extract" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.088139 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6472f81-bf27-4839-a0a4-d7baa206c138" containerName="extract" Nov 25 17:01:28 crc kubenswrapper[4802]: E1125 17:01:28.088164 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6472f81-bf27-4839-a0a4-d7baa206c138" containerName="util" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.088170 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6472f81-bf27-4839-a0a4-d7baa206c138" containerName="util" Nov 25 17:01:28 crc kubenswrapper[4802]: E1125 17:01:28.088178 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6472f81-bf27-4839-a0a4-d7baa206c138" containerName="pull" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.088184 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6472f81-bf27-4839-a0a4-d7baa206c138" containerName="pull" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.088291 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6472f81-bf27-4839-a0a4-d7baa206c138" containerName="extract" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.088955 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.092251 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.092975 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-4d458" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.105640 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf"] Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.164421 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eda71d4e-8d95-4944-a0b8-a8ac84c01ca4-apiservice-cert\") pod \"infra-operator-controller-manager-6fccf5cdb-rrwcf\" (UID: \"eda71d4e-8d95-4944-a0b8-a8ac84c01ca4\") " pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.164491 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6wbt\" (UniqueName: \"kubernetes.io/projected/eda71d4e-8d95-4944-a0b8-a8ac84c01ca4-kube-api-access-l6wbt\") pod \"infra-operator-controller-manager-6fccf5cdb-rrwcf\" (UID: \"eda71d4e-8d95-4944-a0b8-a8ac84c01ca4\") " pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.164583 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eda71d4e-8d95-4944-a0b8-a8ac84c01ca4-webhook-cert\") pod \"infra-operator-controller-manager-6fccf5cdb-rrwcf\" (UID: \"eda71d4e-8d95-4944-a0b8-a8ac84c01ca4\") " pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.265907 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eda71d4e-8d95-4944-a0b8-a8ac84c01ca4-webhook-cert\") pod \"infra-operator-controller-manager-6fccf5cdb-rrwcf\" (UID: \"eda71d4e-8d95-4944-a0b8-a8ac84c01ca4\") " pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.265973 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eda71d4e-8d95-4944-a0b8-a8ac84c01ca4-apiservice-cert\") pod \"infra-operator-controller-manager-6fccf5cdb-rrwcf\" (UID: \"eda71d4e-8d95-4944-a0b8-a8ac84c01ca4\") " pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.266001 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6wbt\" (UniqueName: \"kubernetes.io/projected/eda71d4e-8d95-4944-a0b8-a8ac84c01ca4-kube-api-access-l6wbt\") pod \"infra-operator-controller-manager-6fccf5cdb-rrwcf\" (UID: \"eda71d4e-8d95-4944-a0b8-a8ac84c01ca4\") " pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.272578 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eda71d4e-8d95-4944-a0b8-a8ac84c01ca4-apiservice-cert\") pod \"infra-operator-controller-manager-6fccf5cdb-rrwcf\" (UID: \"eda71d4e-8d95-4944-a0b8-a8ac84c01ca4\") " pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.278845 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eda71d4e-8d95-4944-a0b8-a8ac84c01ca4-webhook-cert\") pod \"infra-operator-controller-manager-6fccf5cdb-rrwcf\" (UID: \"eda71d4e-8d95-4944-a0b8-a8ac84c01ca4\") " pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.284593 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6wbt\" (UniqueName: \"kubernetes.io/projected/eda71d4e-8d95-4944-a0b8-a8ac84c01ca4-kube-api-access-l6wbt\") pod \"infra-operator-controller-manager-6fccf5cdb-rrwcf\" (UID: \"eda71d4e-8d95-4944-a0b8-a8ac84c01ca4\") " pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.407627 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:01:28 crc kubenswrapper[4802]: I1125 17:01:28.800060 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf"] Nov 25 17:01:28 crc kubenswrapper[4802]: W1125 17:01:28.816371 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeda71d4e_8d95_4944_a0b8_a8ac84c01ca4.slice/crio-f6976db09e9a5d051993beca1ae4041125578dab01ca5b605b31480704a8761c WatchSource:0}: Error finding container f6976db09e9a5d051993beca1ae4041125578dab01ca5b605b31480704a8761c: Status 404 returned error can't find the container with id f6976db09e9a5d051993beca1ae4041125578dab01ca5b605b31480704a8761c Nov 25 17:01:29 crc kubenswrapper[4802]: I1125 17:01:29.362412 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" event={"ID":"eda71d4e-8d95-4944-a0b8-a8ac84c01ca4","Type":"ContainerStarted","Data":"f6976db09e9a5d051993beca1ae4041125578dab01ca5b605b31480704a8761c"} Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.305474 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.306710 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.310066 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.310353 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-v2c2s" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.310659 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.311800 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.312033 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.312301 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.318197 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.322703 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.325055 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.348898 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.357075 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.386212 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.509111 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6d23bfa4-e3e1-4942-82d3-dc6d548b0a98-kolla-config\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.509457 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ea5778f-cef3-42cd-b4bb-1fc73a0ee456-operator-scripts\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.509487 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab-config-data-generated\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.509559 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.509762 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3ea5778f-cef3-42cd-b4bb-1fc73a0ee456-config-data-generated\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.509869 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d23bfa4-e3e1-4942-82d3-dc6d548b0a98-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.510003 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22cwh\" (UniqueName: \"kubernetes.io/projected/4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab-kube-api-access-22cwh\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.510206 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab-kolla-config\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.510352 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3ea5778f-cef3-42cd-b4bb-1fc73a0ee456-kolla-config\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.510489 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.510635 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6d23bfa4-e3e1-4942-82d3-dc6d548b0a98-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.510873 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6d23bfa4-e3e1-4942-82d3-dc6d548b0a98-config-data-default\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.511008 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab-config-data-default\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.511154 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3ea5778f-cef3-42cd-b4bb-1fc73a0ee456-config-data-default\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.511268 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.511375 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab-operator-scripts\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.511506 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl5rp\" (UniqueName: \"kubernetes.io/projected/6d23bfa4-e3e1-4942-82d3-dc6d548b0a98-kube-api-access-gl5rp\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.511693 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg8kq\" (UniqueName: \"kubernetes.io/projected/3ea5778f-cef3-42cd-b4bb-1fc73a0ee456-kube-api-access-lg8kq\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.612725 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl5rp\" (UniqueName: \"kubernetes.io/projected/6d23bfa4-e3e1-4942-82d3-dc6d548b0a98-kube-api-access-gl5rp\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.612783 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg8kq\" (UniqueName: \"kubernetes.io/projected/3ea5778f-cef3-42cd-b4bb-1fc73a0ee456-kube-api-access-lg8kq\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.612832 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6d23bfa4-e3e1-4942-82d3-dc6d548b0a98-kolla-config\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.612860 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ea5778f-cef3-42cd-b4bb-1fc73a0ee456-operator-scripts\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.612892 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab-config-data-generated\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.612912 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.612934 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3ea5778f-cef3-42cd-b4bb-1fc73a0ee456-config-data-generated\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.612955 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d23bfa4-e3e1-4942-82d3-dc6d548b0a98-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.612993 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22cwh\" (UniqueName: \"kubernetes.io/projected/4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab-kube-api-access-22cwh\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.613056 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab-kolla-config\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.613083 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3ea5778f-cef3-42cd-b4bb-1fc73a0ee456-kolla-config\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.613112 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.613169 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6d23bfa4-e3e1-4942-82d3-dc6d548b0a98-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.613190 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6d23bfa4-e3e1-4942-82d3-dc6d548b0a98-config-data-default\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.613211 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab-config-data-default\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.613241 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3ea5778f-cef3-42cd-b4bb-1fc73a0ee456-config-data-default\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.613264 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.613285 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab-operator-scripts\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.613504 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.613802 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.614023 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.614285 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6d23bfa4-e3e1-4942-82d3-dc6d548b0a98-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.614392 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3ea5778f-cef3-42cd-b4bb-1fc73a0ee456-config-data-generated\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.614511 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab-config-data-generated\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.614738 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab-kolla-config\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.614831 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab-config-data-default\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.615536 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab-operator-scripts\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.615666 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6d23bfa4-e3e1-4942-82d3-dc6d548b0a98-config-data-default\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.616032 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d23bfa4-e3e1-4942-82d3-dc6d548b0a98-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.616325 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6d23bfa4-e3e1-4942-82d3-dc6d548b0a98-kolla-config\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.618869 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3ea5778f-cef3-42cd-b4bb-1fc73a0ee456-config-data-default\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.618999 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3ea5778f-cef3-42cd-b4bb-1fc73a0ee456-kolla-config\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.619638 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ea5778f-cef3-42cd-b4bb-1fc73a0ee456-operator-scripts\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.631918 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.632460 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl5rp\" (UniqueName: \"kubernetes.io/projected/6d23bfa4-e3e1-4942-82d3-dc6d548b0a98-kube-api-access-gl5rp\") pod \"openstack-galera-0\" (UID: \"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.633404 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg8kq\" (UniqueName: \"kubernetes.io/projected/3ea5778f-cef3-42cd-b4bb-1fc73a0ee456-kube-api-access-lg8kq\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.633485 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.634965 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22cwh\" (UniqueName: \"kubernetes.io/projected/4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab-kube-api-access-22cwh\") pod \"openstack-galera-1\" (UID: \"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.635863 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-2\" (UID: \"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.642984 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.667891 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.681852 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.965646 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 25 17:01:31 crc kubenswrapper[4802]: I1125 17:01:31.988299 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 25 17:01:31 crc kubenswrapper[4802]: W1125 17:01:31.990677 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d23bfa4_e3e1_4942_82d3_dc6d548b0a98.slice/crio-15444997572c9c270e01e950b71300d5d2d070e8a1edc8e19b8519edd39bc964 WatchSource:0}: Error finding container 15444997572c9c270e01e950b71300d5d2d070e8a1edc8e19b8519edd39bc964: Status 404 returned error can't find the container with id 15444997572c9c270e01e950b71300d5d2d070e8a1edc8e19b8519edd39bc964 Nov 25 17:01:32 crc kubenswrapper[4802]: I1125 17:01:32.085509 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 25 17:01:32 crc kubenswrapper[4802]: W1125 17:01:32.089292 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f15a614_5ba1_42d4_a9d5_1daa1a7bf8ab.slice/crio-8211acc51e7364951a19b9f814f7f3368aed72f4832dcc8416366e382b6410fb WatchSource:0}: Error finding container 8211acc51e7364951a19b9f814f7f3368aed72f4832dcc8416366e382b6410fb: Status 404 returned error can't find the container with id 8211acc51e7364951a19b9f814f7f3368aed72f4832dcc8416366e382b6410fb Nov 25 17:01:32 crc kubenswrapper[4802]: I1125 17:01:32.383773 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" event={"ID":"eda71d4e-8d95-4944-a0b8-a8ac84c01ca4","Type":"ContainerStarted","Data":"5fdb255c3eda15bc7a4f2954f1a8b55564ad5bf1940a9e1963595f5ec97e4742"} Nov 25 17:01:32 crc kubenswrapper[4802]: I1125 17:01:32.385184 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456","Type":"ContainerStarted","Data":"d48769b5338242e613ea03ddafbb801b42ff6b9daada411ae4c5d71c6ab55c6b"} Nov 25 17:01:32 crc kubenswrapper[4802]: I1125 17:01:32.386136 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98","Type":"ContainerStarted","Data":"15444997572c9c270e01e950b71300d5d2d070e8a1edc8e19b8519edd39bc964"} Nov 25 17:01:32 crc kubenswrapper[4802]: I1125 17:01:32.387255 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab","Type":"ContainerStarted","Data":"8211acc51e7364951a19b9f814f7f3368aed72f4832dcc8416366e382b6410fb"} Nov 25 17:01:35 crc kubenswrapper[4802]: I1125 17:01:35.411959 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" event={"ID":"eda71d4e-8d95-4944-a0b8-a8ac84c01ca4","Type":"ContainerStarted","Data":"5d22275b4f74969967fed7b7141489a59b7aaf3161809c6c06a7a14f6ecddba4"} Nov 25 17:01:35 crc kubenswrapper[4802]: I1125 17:01:35.412495 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:01:35 crc kubenswrapper[4802]: I1125 17:01:35.437723 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" podStartSLOduration=0.997645057 podStartE2EDuration="7.437699075s" podCreationTimestamp="2025-11-25 17:01:28 +0000 UTC" firstStartedPulling="2025-11-25 17:01:28.818740322 +0000 UTC m=+871.963087508" lastFinishedPulling="2025-11-25 17:01:35.25879434 +0000 UTC m=+878.403141526" observedRunningTime="2025-11-25 17:01:35.437016477 +0000 UTC m=+878.581363663" watchObservedRunningTime="2025-11-25 17:01:35.437699075 +0000 UTC m=+878.582046261" Nov 25 17:01:36 crc kubenswrapper[4802]: I1125 17:01:36.425377 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:01:37 crc kubenswrapper[4802]: I1125 17:01:37.351676 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 25 17:01:37 crc kubenswrapper[4802]: I1125 17:01:37.352916 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 25 17:01:37 crc kubenswrapper[4802]: I1125 17:01:37.354823 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-7zbkx" Nov 25 17:01:37 crc kubenswrapper[4802]: I1125 17:01:37.355167 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Nov 25 17:01:37 crc kubenswrapper[4802]: I1125 17:01:37.368084 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 25 17:01:37 crc kubenswrapper[4802]: I1125 17:01:37.508590 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8357d6cc-810d-42bc-b387-f3972fede479-config-data\") pod \"memcached-0\" (UID: \"8357d6cc-810d-42bc-b387-f3972fede479\") " pod="glance-kuttl-tests/memcached-0" Nov 25 17:01:37 crc kubenswrapper[4802]: I1125 17:01:37.508674 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8357d6cc-810d-42bc-b387-f3972fede479-kolla-config\") pod \"memcached-0\" (UID: \"8357d6cc-810d-42bc-b387-f3972fede479\") " pod="glance-kuttl-tests/memcached-0" Nov 25 17:01:37 crc kubenswrapper[4802]: I1125 17:01:37.508706 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2h9k\" (UniqueName: \"kubernetes.io/projected/8357d6cc-810d-42bc-b387-f3972fede479-kube-api-access-c2h9k\") pod \"memcached-0\" (UID: \"8357d6cc-810d-42bc-b387-f3972fede479\") " pod="glance-kuttl-tests/memcached-0" Nov 25 17:01:37 crc kubenswrapper[4802]: I1125 17:01:37.610504 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8357d6cc-810d-42bc-b387-f3972fede479-config-data\") pod \"memcached-0\" (UID: \"8357d6cc-810d-42bc-b387-f3972fede479\") " pod="glance-kuttl-tests/memcached-0" Nov 25 17:01:37 crc kubenswrapper[4802]: I1125 17:01:37.610603 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8357d6cc-810d-42bc-b387-f3972fede479-kolla-config\") pod \"memcached-0\" (UID: \"8357d6cc-810d-42bc-b387-f3972fede479\") " pod="glance-kuttl-tests/memcached-0" Nov 25 17:01:37 crc kubenswrapper[4802]: I1125 17:01:37.610630 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2h9k\" (UniqueName: \"kubernetes.io/projected/8357d6cc-810d-42bc-b387-f3972fede479-kube-api-access-c2h9k\") pod \"memcached-0\" (UID: \"8357d6cc-810d-42bc-b387-f3972fede479\") " pod="glance-kuttl-tests/memcached-0" Nov 25 17:01:37 crc kubenswrapper[4802]: I1125 17:01:37.611525 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8357d6cc-810d-42bc-b387-f3972fede479-config-data\") pod \"memcached-0\" (UID: \"8357d6cc-810d-42bc-b387-f3972fede479\") " pod="glance-kuttl-tests/memcached-0" Nov 25 17:01:37 crc kubenswrapper[4802]: I1125 17:01:37.611592 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8357d6cc-810d-42bc-b387-f3972fede479-kolla-config\") pod \"memcached-0\" (UID: \"8357d6cc-810d-42bc-b387-f3972fede479\") " pod="glance-kuttl-tests/memcached-0" Nov 25 17:01:37 crc kubenswrapper[4802]: I1125 17:01:37.631204 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2h9k\" (UniqueName: \"kubernetes.io/projected/8357d6cc-810d-42bc-b387-f3972fede479-kube-api-access-c2h9k\") pod \"memcached-0\" (UID: \"8357d6cc-810d-42bc-b387-f3972fede479\") " pod="glance-kuttl-tests/memcached-0" Nov 25 17:01:37 crc kubenswrapper[4802]: I1125 17:01:37.668493 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 25 17:01:40 crc kubenswrapper[4802]: I1125 17:01:40.301688 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-fcpts"] Nov 25 17:01:40 crc kubenswrapper[4802]: I1125 17:01:40.304427 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-fcpts" Nov 25 17:01:40 crc kubenswrapper[4802]: I1125 17:01:40.305173 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-fcpts"] Nov 25 17:01:40 crc kubenswrapper[4802]: I1125 17:01:40.307387 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-jfwbm" Nov 25 17:01:40 crc kubenswrapper[4802]: I1125 17:01:40.453256 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lns7z\" (UniqueName: \"kubernetes.io/projected/951981e9-7b86-4a70-859c-d6cdcac32468-kube-api-access-lns7z\") pod \"rabbitmq-cluster-operator-index-fcpts\" (UID: \"951981e9-7b86-4a70-859c-d6cdcac32468\") " pod="openstack-operators/rabbitmq-cluster-operator-index-fcpts" Nov 25 17:01:40 crc kubenswrapper[4802]: I1125 17:01:40.554476 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lns7z\" (UniqueName: \"kubernetes.io/projected/951981e9-7b86-4a70-859c-d6cdcac32468-kube-api-access-lns7z\") pod \"rabbitmq-cluster-operator-index-fcpts\" (UID: \"951981e9-7b86-4a70-859c-d6cdcac32468\") " pod="openstack-operators/rabbitmq-cluster-operator-index-fcpts" Nov 25 17:01:40 crc kubenswrapper[4802]: I1125 17:01:40.574781 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lns7z\" (UniqueName: \"kubernetes.io/projected/951981e9-7b86-4a70-859c-d6cdcac32468-kube-api-access-lns7z\") pod \"rabbitmq-cluster-operator-index-fcpts\" (UID: \"951981e9-7b86-4a70-859c-d6cdcac32468\") " pod="openstack-operators/rabbitmq-cluster-operator-index-fcpts" Nov 25 17:01:40 crc kubenswrapper[4802]: I1125 17:01:40.628654 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-fcpts" Nov 25 17:01:42 crc kubenswrapper[4802]: I1125 17:01:42.236163 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 25 17:01:42 crc kubenswrapper[4802]: I1125 17:01:42.289710 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-fcpts"] Nov 25 17:01:42 crc kubenswrapper[4802]: W1125 17:01:42.295328 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod951981e9_7b86_4a70_859c_d6cdcac32468.slice/crio-05d17b2a5771759c8d2cb36292a70f0472b794302d54b988c8f24d9d5c4556c3 WatchSource:0}: Error finding container 05d17b2a5771759c8d2cb36292a70f0472b794302d54b988c8f24d9d5c4556c3: Status 404 returned error can't find the container with id 05d17b2a5771759c8d2cb36292a70f0472b794302d54b988c8f24d9d5c4556c3 Nov 25 17:01:42 crc kubenswrapper[4802]: I1125 17:01:42.459556 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"8357d6cc-810d-42bc-b387-f3972fede479","Type":"ContainerStarted","Data":"1eee514fad8ce2f941a4fe63d0186f098847e26b0dd6371bd441eacfef62cdb4"} Nov 25 17:01:42 crc kubenswrapper[4802]: I1125 17:01:42.460650 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-fcpts" event={"ID":"951981e9-7b86-4a70-859c-d6cdcac32468","Type":"ContainerStarted","Data":"05d17b2a5771759c8d2cb36292a70f0472b794302d54b988c8f24d9d5c4556c3"} Nov 25 17:01:44 crc kubenswrapper[4802]: I1125 17:01:43.469701 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98","Type":"ContainerStarted","Data":"65cbf228b5c0699584f5308808b91b0865134baeff240535813bc2c8b97e4bca"} Nov 25 17:01:44 crc kubenswrapper[4802]: I1125 17:01:43.472726 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab","Type":"ContainerStarted","Data":"7c07a70c4666b8bcc55b51cc3c561cc5a0df9b69d69e467d05f51b18a7e91f5b"} Nov 25 17:01:45 crc kubenswrapper[4802]: I1125 17:01:45.090472 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-fcpts"] Nov 25 17:01:45 crc kubenswrapper[4802]: I1125 17:01:45.485997 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456","Type":"ContainerStarted","Data":"a32810f672acb3fba4053352a3efab6fd8dfc6bd8e365703cfd199cc8e996019"} Nov 25 17:01:45 crc kubenswrapper[4802]: I1125 17:01:45.691811 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-984qp"] Nov 25 17:01:45 crc kubenswrapper[4802]: I1125 17:01:45.692657 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-984qp" Nov 25 17:01:45 crc kubenswrapper[4802]: I1125 17:01:45.707626 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-984qp"] Nov 25 17:01:45 crc kubenswrapper[4802]: I1125 17:01:45.829152 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ldn5\" (UniqueName: \"kubernetes.io/projected/7fec300a-5625-4876-b0a0-d82c18409f4b-kube-api-access-5ldn5\") pod \"rabbitmq-cluster-operator-index-984qp\" (UID: \"7fec300a-5625-4876-b0a0-d82c18409f4b\") " pod="openstack-operators/rabbitmq-cluster-operator-index-984qp" Nov 25 17:01:45 crc kubenswrapper[4802]: I1125 17:01:45.931047 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ldn5\" (UniqueName: \"kubernetes.io/projected/7fec300a-5625-4876-b0a0-d82c18409f4b-kube-api-access-5ldn5\") pod \"rabbitmq-cluster-operator-index-984qp\" (UID: \"7fec300a-5625-4876-b0a0-d82c18409f4b\") " pod="openstack-operators/rabbitmq-cluster-operator-index-984qp" Nov 25 17:01:45 crc kubenswrapper[4802]: I1125 17:01:45.952528 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ldn5\" (UniqueName: \"kubernetes.io/projected/7fec300a-5625-4876-b0a0-d82c18409f4b-kube-api-access-5ldn5\") pod \"rabbitmq-cluster-operator-index-984qp\" (UID: \"7fec300a-5625-4876-b0a0-d82c18409f4b\") " pod="openstack-operators/rabbitmq-cluster-operator-index-984qp" Nov 25 17:01:46 crc kubenswrapper[4802]: I1125 17:01:46.007666 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-984qp" Nov 25 17:01:46 crc kubenswrapper[4802]: I1125 17:01:46.280061 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-984qp"] Nov 25 17:01:46 crc kubenswrapper[4802]: W1125 17:01:46.290847 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fec300a_5625_4876_b0a0_d82c18409f4b.slice/crio-98cf82bcf385f27d783b13a107dfebfce68455e7688a75abafed6399f90824f7 WatchSource:0}: Error finding container 98cf82bcf385f27d783b13a107dfebfce68455e7688a75abafed6399f90824f7: Status 404 returned error can't find the container with id 98cf82bcf385f27d783b13a107dfebfce68455e7688a75abafed6399f90824f7 Nov 25 17:01:46 crc kubenswrapper[4802]: I1125 17:01:46.492581 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-984qp" event={"ID":"7fec300a-5625-4876-b0a0-d82c18409f4b","Type":"ContainerStarted","Data":"98cf82bcf385f27d783b13a107dfebfce68455e7688a75abafed6399f90824f7"} Nov 25 17:02:05 crc kubenswrapper[4802]: E1125 17:02:05.419924 4802 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator-index:latest" Nov 25 17:02:05 crc kubenswrapper[4802]: E1125 17:02:05.420602 4802 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5ldn5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-index-984qp_openstack-operators(7fec300a-5625-4876-b0a0-d82c18409f4b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 17:02:05 crc kubenswrapper[4802]: E1125 17:02:05.422087 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-index-984qp" podUID="7fec300a-5625-4876-b0a0-d82c18409f4b" Nov 25 17:02:05 crc kubenswrapper[4802]: I1125 17:02:05.609333 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"8357d6cc-810d-42bc-b387-f3972fede479","Type":"ContainerStarted","Data":"c5cd062a95ae35dc883c8a32356dbd61d89e5ef7c6314c04a9e72600874a4501"} Nov 25 17:02:05 crc kubenswrapper[4802]: I1125 17:02:05.632896 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/memcached-0" podStartSLOduration=11.497442448 podStartE2EDuration="28.632868539s" podCreationTimestamp="2025-11-25 17:01:37 +0000 UTC" firstStartedPulling="2025-11-25 17:01:42.254899892 +0000 UTC m=+885.399247088" lastFinishedPulling="2025-11-25 17:01:59.390325983 +0000 UTC m=+902.534673179" observedRunningTime="2025-11-25 17:02:05.628227915 +0000 UTC m=+908.772575111" watchObservedRunningTime="2025-11-25 17:02:05.632868539 +0000 UTC m=+908.777215755" Nov 25 17:02:05 crc kubenswrapper[4802]: E1125 17:02:05.918789 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator-index:latest\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-index-984qp" podUID="7fec300a-5625-4876-b0a0-d82c18409f4b" Nov 25 17:02:06 crc kubenswrapper[4802]: I1125 17:02:06.618545 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-fcpts" event={"ID":"951981e9-7b86-4a70-859c-d6cdcac32468","Type":"ContainerStarted","Data":"a6670c19ca724051a344fdac7c380a9a870f8be52c545f31c258c9f2ae847d12"} Nov 25 17:02:06 crc kubenswrapper[4802]: I1125 17:02:06.618690 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/memcached-0" Nov 25 17:02:06 crc kubenswrapper[4802]: I1125 17:02:06.618803 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-fcpts" podUID="951981e9-7b86-4a70-859c-d6cdcac32468" containerName="registry-server" containerID="cri-o://a6670c19ca724051a344fdac7c380a9a870f8be52c545f31c258c9f2ae847d12" gracePeriod=2 Nov 25 17:02:06 crc kubenswrapper[4802]: I1125 17:02:06.638166 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-fcpts" podStartSLOduration=3.00629853 podStartE2EDuration="26.638110964s" podCreationTimestamp="2025-11-25 17:01:40 +0000 UTC" firstStartedPulling="2025-11-25 17:01:42.297726352 +0000 UTC m=+885.442073538" lastFinishedPulling="2025-11-25 17:02:05.929538786 +0000 UTC m=+909.073885972" observedRunningTime="2025-11-25 17:02:06.635680238 +0000 UTC m=+909.780027434" watchObservedRunningTime="2025-11-25 17:02:06.638110964 +0000 UTC m=+909.782458150" Nov 25 17:02:07 crc kubenswrapper[4802]: I1125 17:02:06.999925 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-fcpts" Nov 25 17:02:07 crc kubenswrapper[4802]: I1125 17:02:07.043184 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lns7z\" (UniqueName: \"kubernetes.io/projected/951981e9-7b86-4a70-859c-d6cdcac32468-kube-api-access-lns7z\") pod \"951981e9-7b86-4a70-859c-d6cdcac32468\" (UID: \"951981e9-7b86-4a70-859c-d6cdcac32468\") " Nov 25 17:02:07 crc kubenswrapper[4802]: I1125 17:02:07.048875 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/951981e9-7b86-4a70-859c-d6cdcac32468-kube-api-access-lns7z" (OuterVolumeSpecName: "kube-api-access-lns7z") pod "951981e9-7b86-4a70-859c-d6cdcac32468" (UID: "951981e9-7b86-4a70-859c-d6cdcac32468"). InnerVolumeSpecName "kube-api-access-lns7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:02:07 crc kubenswrapper[4802]: I1125 17:02:07.144958 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lns7z\" (UniqueName: \"kubernetes.io/projected/951981e9-7b86-4a70-859c-d6cdcac32468-kube-api-access-lns7z\") on node \"crc\" DevicePath \"\"" Nov 25 17:02:07 crc kubenswrapper[4802]: I1125 17:02:07.625810 4802 generic.go:334] "Generic (PLEG): container finished" podID="4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab" containerID="7c07a70c4666b8bcc55b51cc3c561cc5a0df9b69d69e467d05f51b18a7e91f5b" exitCode=0 Nov 25 17:02:07 crc kubenswrapper[4802]: I1125 17:02:07.625898 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab","Type":"ContainerDied","Data":"7c07a70c4666b8bcc55b51cc3c561cc5a0df9b69d69e467d05f51b18a7e91f5b"} Nov 25 17:02:07 crc kubenswrapper[4802]: I1125 17:02:07.627431 4802 generic.go:334] "Generic (PLEG): container finished" podID="951981e9-7b86-4a70-859c-d6cdcac32468" containerID="a6670c19ca724051a344fdac7c380a9a870f8be52c545f31c258c9f2ae847d12" exitCode=0 Nov 25 17:02:07 crc kubenswrapper[4802]: I1125 17:02:07.627724 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-fcpts" event={"ID":"951981e9-7b86-4a70-859c-d6cdcac32468","Type":"ContainerDied","Data":"a6670c19ca724051a344fdac7c380a9a870f8be52c545f31c258c9f2ae847d12"} Nov 25 17:02:07 crc kubenswrapper[4802]: I1125 17:02:07.627794 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-fcpts" event={"ID":"951981e9-7b86-4a70-859c-d6cdcac32468","Type":"ContainerDied","Data":"05d17b2a5771759c8d2cb36292a70f0472b794302d54b988c8f24d9d5c4556c3"} Nov 25 17:02:07 crc kubenswrapper[4802]: I1125 17:02:07.627827 4802 scope.go:117] "RemoveContainer" containerID="a6670c19ca724051a344fdac7c380a9a870f8be52c545f31c258c9f2ae847d12" Nov 25 17:02:07 crc kubenswrapper[4802]: I1125 17:02:07.627994 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-fcpts" Nov 25 17:02:07 crc kubenswrapper[4802]: I1125 17:02:07.675950 4802 scope.go:117] "RemoveContainer" containerID="a6670c19ca724051a344fdac7c380a9a870f8be52c545f31c258c9f2ae847d12" Nov 25 17:02:07 crc kubenswrapper[4802]: E1125 17:02:07.676734 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6670c19ca724051a344fdac7c380a9a870f8be52c545f31c258c9f2ae847d12\": container with ID starting with a6670c19ca724051a344fdac7c380a9a870f8be52c545f31c258c9f2ae847d12 not found: ID does not exist" containerID="a6670c19ca724051a344fdac7c380a9a870f8be52c545f31c258c9f2ae847d12" Nov 25 17:02:07 crc kubenswrapper[4802]: I1125 17:02:07.676778 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6670c19ca724051a344fdac7c380a9a870f8be52c545f31c258c9f2ae847d12"} err="failed to get container status \"a6670c19ca724051a344fdac7c380a9a870f8be52c545f31c258c9f2ae847d12\": rpc error: code = NotFound desc = could not find container \"a6670c19ca724051a344fdac7c380a9a870f8be52c545f31c258c9f2ae847d12\": container with ID starting with a6670c19ca724051a344fdac7c380a9a870f8be52c545f31c258c9f2ae847d12 not found: ID does not exist" Nov 25 17:02:07 crc kubenswrapper[4802]: I1125 17:02:07.680900 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-fcpts"] Nov 25 17:02:07 crc kubenswrapper[4802]: I1125 17:02:07.686340 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-fcpts"] Nov 25 17:02:08 crc kubenswrapper[4802]: I1125 17:02:08.635429 4802 generic.go:334] "Generic (PLEG): container finished" podID="3ea5778f-cef3-42cd-b4bb-1fc73a0ee456" containerID="a32810f672acb3fba4053352a3efab6fd8dfc6bd8e365703cfd199cc8e996019" exitCode=0 Nov 25 17:02:08 crc kubenswrapper[4802]: I1125 17:02:08.635494 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456","Type":"ContainerDied","Data":"a32810f672acb3fba4053352a3efab6fd8dfc6bd8e365703cfd199cc8e996019"} Nov 25 17:02:08 crc kubenswrapper[4802]: I1125 17:02:08.637652 4802 generic.go:334] "Generic (PLEG): container finished" podID="6d23bfa4-e3e1-4942-82d3-dc6d548b0a98" containerID="65cbf228b5c0699584f5308808b91b0865134baeff240535813bc2c8b97e4bca" exitCode=0 Nov 25 17:02:08 crc kubenswrapper[4802]: I1125 17:02:08.637705 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98","Type":"ContainerDied","Data":"65cbf228b5c0699584f5308808b91b0865134baeff240535813bc2c8b97e4bca"} Nov 25 17:02:08 crc kubenswrapper[4802]: I1125 17:02:08.640005 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab","Type":"ContainerStarted","Data":"c5976d483ae7e45b19f8888cf9b9a89533d92b73711c12fc4917908b6cfaaddd"} Nov 25 17:02:08 crc kubenswrapper[4802]: I1125 17:02:08.679166 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-1" podStartSLOduration=28.133380581 podStartE2EDuration="38.679149693s" podCreationTimestamp="2025-11-25 17:01:30 +0000 UTC" firstStartedPulling="2025-11-25 17:01:32.091635061 +0000 UTC m=+875.235982247" lastFinishedPulling="2025-11-25 17:01:42.637404173 +0000 UTC m=+885.781751359" observedRunningTime="2025-11-25 17:02:08.675028692 +0000 UTC m=+911.819375888" watchObservedRunningTime="2025-11-25 17:02:08.679149693 +0000 UTC m=+911.823496869" Nov 25 17:02:09 crc kubenswrapper[4802]: I1125 17:02:09.513655 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="951981e9-7b86-4a70-859c-d6cdcac32468" path="/var/lib/kubelet/pods/951981e9-7b86-4a70-859c-d6cdcac32468/volumes" Nov 25 17:02:09 crc kubenswrapper[4802]: I1125 17:02:09.651555 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"3ea5778f-cef3-42cd-b4bb-1fc73a0ee456","Type":"ContainerStarted","Data":"62e30c6dc50eae7826655763d7c55e9078d9f4cf20f7f4fab79e37b6fed057af"} Nov 25 17:02:09 crc kubenswrapper[4802]: I1125 17:02:09.654347 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"6d23bfa4-e3e1-4942-82d3-dc6d548b0a98","Type":"ContainerStarted","Data":"c3f54b8e09377616af4e0f9fcd94a5ea28ddf4e32ba3f418a11fe3aebd0781d5"} Nov 25 17:02:11 crc kubenswrapper[4802]: I1125 17:02:11.644375 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:02:11 crc kubenswrapper[4802]: I1125 17:02:11.644673 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:02:12 crc kubenswrapper[4802]: I1125 17:02:12.669760 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/memcached-0" Nov 25 17:02:14 crc kubenswrapper[4802]: I1125 17:02:14.705150 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-0" podStartSLOduration=33.744943392 podStartE2EDuration="44.705104362s" podCreationTimestamp="2025-11-25 17:01:30 +0000 UTC" firstStartedPulling="2025-11-25 17:01:31.997023251 +0000 UTC m=+875.141370437" lastFinishedPulling="2025-11-25 17:01:42.957184221 +0000 UTC m=+886.101531407" observedRunningTime="2025-11-25 17:02:14.703005416 +0000 UTC m=+917.847352602" watchObservedRunningTime="2025-11-25 17:02:14.705104362 +0000 UTC m=+917.849451548" Nov 25 17:02:14 crc kubenswrapper[4802]: I1125 17:02:14.724468 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-2" podStartSLOduration=31.921255938 podStartE2EDuration="44.724451421s" podCreationTimestamp="2025-11-25 17:01:30 +0000 UTC" firstStartedPulling="2025-11-25 17:01:31.966021758 +0000 UTC m=+875.110368944" lastFinishedPulling="2025-11-25 17:01:44.769217241 +0000 UTC m=+887.913564427" observedRunningTime="2025-11-25 17:02:14.72104953 +0000 UTC m=+917.865396726" watchObservedRunningTime="2025-11-25 17:02:14.724451421 +0000 UTC m=+917.868798607" Nov 25 17:02:17 crc kubenswrapper[4802]: I1125 17:02:17.705181 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-984qp" event={"ID":"7fec300a-5625-4876-b0a0-d82c18409f4b","Type":"ContainerStarted","Data":"c0b825131415930d249b02aa82d865600bd0ff3ad91152ca37b361a53e601a2b"} Nov 25 17:02:17 crc kubenswrapper[4802]: I1125 17:02:17.725046 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-984qp" podStartSLOduration=2.114923312 podStartE2EDuration="32.725028048s" podCreationTimestamp="2025-11-25 17:01:45 +0000 UTC" firstStartedPulling="2025-11-25 17:01:46.294082309 +0000 UTC m=+889.438429495" lastFinishedPulling="2025-11-25 17:02:16.904187045 +0000 UTC m=+920.048534231" observedRunningTime="2025-11-25 17:02:17.721320969 +0000 UTC m=+920.865668155" watchObservedRunningTime="2025-11-25 17:02:17.725028048 +0000 UTC m=+920.869375234" Nov 25 17:02:21 crc kubenswrapper[4802]: I1125 17:02:21.668704 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:02:21 crc kubenswrapper[4802]: I1125 17:02:21.670429 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:02:21 crc kubenswrapper[4802]: I1125 17:02:21.682033 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:02:21 crc kubenswrapper[4802]: I1125 17:02:21.682085 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:02:21 crc kubenswrapper[4802]: I1125 17:02:21.731560 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:02:21 crc kubenswrapper[4802]: I1125 17:02:21.794903 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 17:02:24 crc kubenswrapper[4802]: I1125 17:02:24.248754 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 17:02:24 crc kubenswrapper[4802]: I1125 17:02:24.249626 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 17:02:26 crc kubenswrapper[4802]: I1125 17:02:26.008507 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-984qp" Nov 25 17:02:26 crc kubenswrapper[4802]: I1125 17:02:26.008771 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-984qp" Nov 25 17:02:26 crc kubenswrapper[4802]: I1125 17:02:26.040591 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-984qp" Nov 25 17:02:26 crc kubenswrapper[4802]: I1125 17:02:26.779824 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-984qp" Nov 25 17:02:30 crc kubenswrapper[4802]: I1125 17:02:30.484783 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:02:30 crc kubenswrapper[4802]: I1125 17:02:30.547871 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 17:02:30 crc kubenswrapper[4802]: I1125 17:02:30.739180 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc"] Nov 25 17:02:30 crc kubenswrapper[4802]: E1125 17:02:30.739529 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="951981e9-7b86-4a70-859c-d6cdcac32468" containerName="registry-server" Nov 25 17:02:30 crc kubenswrapper[4802]: I1125 17:02:30.739556 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="951981e9-7b86-4a70-859c-d6cdcac32468" containerName="registry-server" Nov 25 17:02:30 crc kubenswrapper[4802]: I1125 17:02:30.739735 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="951981e9-7b86-4a70-859c-d6cdcac32468" containerName="registry-server" Nov 25 17:02:30 crc kubenswrapper[4802]: I1125 17:02:30.740827 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc" Nov 25 17:02:30 crc kubenswrapper[4802]: I1125 17:02:30.743211 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-t7662" Nov 25 17:02:30 crc kubenswrapper[4802]: I1125 17:02:30.749976 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc"] Nov 25 17:02:30 crc kubenswrapper[4802]: I1125 17:02:30.885563 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/376dff9f-509e-46c3-b110-8bf2c43dffda-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc\" (UID: \"376dff9f-509e-46c3-b110-8bf2c43dffda\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc" Nov 25 17:02:30 crc kubenswrapper[4802]: I1125 17:02:30.885604 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47lj5\" (UniqueName: \"kubernetes.io/projected/376dff9f-509e-46c3-b110-8bf2c43dffda-kube-api-access-47lj5\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc\" (UID: \"376dff9f-509e-46c3-b110-8bf2c43dffda\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc" Nov 25 17:02:30 crc kubenswrapper[4802]: I1125 17:02:30.885847 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/376dff9f-509e-46c3-b110-8bf2c43dffda-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc\" (UID: \"376dff9f-509e-46c3-b110-8bf2c43dffda\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc" Nov 25 17:02:30 crc kubenswrapper[4802]: I1125 17:02:30.987396 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/376dff9f-509e-46c3-b110-8bf2c43dffda-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc\" (UID: \"376dff9f-509e-46c3-b110-8bf2c43dffda\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc" Nov 25 17:02:30 crc kubenswrapper[4802]: I1125 17:02:30.987462 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/376dff9f-509e-46c3-b110-8bf2c43dffda-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc\" (UID: \"376dff9f-509e-46c3-b110-8bf2c43dffda\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc" Nov 25 17:02:30 crc kubenswrapper[4802]: I1125 17:02:30.987486 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47lj5\" (UniqueName: \"kubernetes.io/projected/376dff9f-509e-46c3-b110-8bf2c43dffda-kube-api-access-47lj5\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc\" (UID: \"376dff9f-509e-46c3-b110-8bf2c43dffda\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc" Nov 25 17:02:30 crc kubenswrapper[4802]: I1125 17:02:30.988066 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/376dff9f-509e-46c3-b110-8bf2c43dffda-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc\" (UID: \"376dff9f-509e-46c3-b110-8bf2c43dffda\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc" Nov 25 17:02:30 crc kubenswrapper[4802]: I1125 17:02:30.988102 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/376dff9f-509e-46c3-b110-8bf2c43dffda-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc\" (UID: \"376dff9f-509e-46c3-b110-8bf2c43dffda\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc" Nov 25 17:02:31 crc kubenswrapper[4802]: I1125 17:02:31.017378 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47lj5\" (UniqueName: \"kubernetes.io/projected/376dff9f-509e-46c3-b110-8bf2c43dffda-kube-api-access-47lj5\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc\" (UID: \"376dff9f-509e-46c3-b110-8bf2c43dffda\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc" Nov 25 17:02:31 crc kubenswrapper[4802]: I1125 17:02:31.061182 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc" Nov 25 17:02:31 crc kubenswrapper[4802]: I1125 17:02:31.481471 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc"] Nov 25 17:02:31 crc kubenswrapper[4802]: W1125 17:02:31.487396 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod376dff9f_509e_46c3_b110_8bf2c43dffda.slice/crio-18b66db8ec323043c53381b54e0ab132203067d2a2bffeeea5d2cc1f2cc3b534 WatchSource:0}: Error finding container 18b66db8ec323043c53381b54e0ab132203067d2a2bffeeea5d2cc1f2cc3b534: Status 404 returned error can't find the container with id 18b66db8ec323043c53381b54e0ab132203067d2a2bffeeea5d2cc1f2cc3b534 Nov 25 17:02:31 crc kubenswrapper[4802]: I1125 17:02:31.787512 4802 generic.go:334] "Generic (PLEG): container finished" podID="376dff9f-509e-46c3-b110-8bf2c43dffda" containerID="d455b1ca3d670dd594759f1c1e9b99b934e0f7c7d3867bc6ea3765193b62839b" exitCode=0 Nov 25 17:02:31 crc kubenswrapper[4802]: I1125 17:02:31.787560 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc" event={"ID":"376dff9f-509e-46c3-b110-8bf2c43dffda","Type":"ContainerDied","Data":"d455b1ca3d670dd594759f1c1e9b99b934e0f7c7d3867bc6ea3765193b62839b"} Nov 25 17:02:31 crc kubenswrapper[4802]: I1125 17:02:31.787587 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc" event={"ID":"376dff9f-509e-46c3-b110-8bf2c43dffda","Type":"ContainerStarted","Data":"18b66db8ec323043c53381b54e0ab132203067d2a2bffeeea5d2cc1f2cc3b534"} Nov 25 17:02:32 crc kubenswrapper[4802]: I1125 17:02:32.795779 4802 generic.go:334] "Generic (PLEG): container finished" podID="376dff9f-509e-46c3-b110-8bf2c43dffda" containerID="dc2bd83ced3d12c93d93e0091c1d9d0da1f70c1c7ddc086b0bd983d336abe97b" exitCode=0 Nov 25 17:02:32 crc kubenswrapper[4802]: I1125 17:02:32.795944 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc" event={"ID":"376dff9f-509e-46c3-b110-8bf2c43dffda","Type":"ContainerDied","Data":"dc2bd83ced3d12c93d93e0091c1d9d0da1f70c1c7ddc086b0bd983d336abe97b"} Nov 25 17:02:33 crc kubenswrapper[4802]: I1125 17:02:33.665237 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:02:33 crc kubenswrapper[4802]: I1125 17:02:33.743742 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 17:02:33 crc kubenswrapper[4802]: I1125 17:02:33.803603 4802 generic.go:334] "Generic (PLEG): container finished" podID="376dff9f-509e-46c3-b110-8bf2c43dffda" containerID="faa7cf7dd258af11545803f5595a0e9a26d2944097c8a582133c86fe608fe184" exitCode=0 Nov 25 17:02:33 crc kubenswrapper[4802]: I1125 17:02:33.803679 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc" event={"ID":"376dff9f-509e-46c3-b110-8bf2c43dffda","Type":"ContainerDied","Data":"faa7cf7dd258af11545803f5595a0e9a26d2944097c8a582133c86fe608fe184"} Nov 25 17:02:35 crc kubenswrapper[4802]: I1125 17:02:35.061706 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc" Nov 25 17:02:35 crc kubenswrapper[4802]: I1125 17:02:35.145423 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/376dff9f-509e-46c3-b110-8bf2c43dffda-util\") pod \"376dff9f-509e-46c3-b110-8bf2c43dffda\" (UID: \"376dff9f-509e-46c3-b110-8bf2c43dffda\") " Nov 25 17:02:35 crc kubenswrapper[4802]: I1125 17:02:35.145527 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47lj5\" (UniqueName: \"kubernetes.io/projected/376dff9f-509e-46c3-b110-8bf2c43dffda-kube-api-access-47lj5\") pod \"376dff9f-509e-46c3-b110-8bf2c43dffda\" (UID: \"376dff9f-509e-46c3-b110-8bf2c43dffda\") " Nov 25 17:02:35 crc kubenswrapper[4802]: I1125 17:02:35.145579 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/376dff9f-509e-46c3-b110-8bf2c43dffda-bundle\") pod \"376dff9f-509e-46c3-b110-8bf2c43dffda\" (UID: \"376dff9f-509e-46c3-b110-8bf2c43dffda\") " Nov 25 17:02:35 crc kubenswrapper[4802]: I1125 17:02:35.146527 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/376dff9f-509e-46c3-b110-8bf2c43dffda-bundle" (OuterVolumeSpecName: "bundle") pod "376dff9f-509e-46c3-b110-8bf2c43dffda" (UID: "376dff9f-509e-46c3-b110-8bf2c43dffda"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:02:35 crc kubenswrapper[4802]: I1125 17:02:35.150724 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/376dff9f-509e-46c3-b110-8bf2c43dffda-kube-api-access-47lj5" (OuterVolumeSpecName: "kube-api-access-47lj5") pod "376dff9f-509e-46c3-b110-8bf2c43dffda" (UID: "376dff9f-509e-46c3-b110-8bf2c43dffda"). InnerVolumeSpecName "kube-api-access-47lj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:02:35 crc kubenswrapper[4802]: I1125 17:02:35.160783 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/376dff9f-509e-46c3-b110-8bf2c43dffda-util" (OuterVolumeSpecName: "util") pod "376dff9f-509e-46c3-b110-8bf2c43dffda" (UID: "376dff9f-509e-46c3-b110-8bf2c43dffda"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:02:35 crc kubenswrapper[4802]: I1125 17:02:35.246909 4802 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/376dff9f-509e-46c3-b110-8bf2c43dffda-util\") on node \"crc\" DevicePath \"\"" Nov 25 17:02:35 crc kubenswrapper[4802]: I1125 17:02:35.246963 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47lj5\" (UniqueName: \"kubernetes.io/projected/376dff9f-509e-46c3-b110-8bf2c43dffda-kube-api-access-47lj5\") on node \"crc\" DevicePath \"\"" Nov 25 17:02:35 crc kubenswrapper[4802]: I1125 17:02:35.246984 4802 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/376dff9f-509e-46c3-b110-8bf2c43dffda-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 17:02:35 crc kubenswrapper[4802]: I1125 17:02:35.820813 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc" event={"ID":"376dff9f-509e-46c3-b110-8bf2c43dffda","Type":"ContainerDied","Data":"18b66db8ec323043c53381b54e0ab132203067d2a2bffeeea5d2cc1f2cc3b534"} Nov 25 17:02:35 crc kubenswrapper[4802]: I1125 17:02:35.820858 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18b66db8ec323043c53381b54e0ab132203067d2a2bffeeea5d2cc1f2cc3b534" Nov 25 17:02:35 crc kubenswrapper[4802]: I1125 17:02:35.820925 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc" Nov 25 17:02:40 crc kubenswrapper[4802]: I1125 17:02:40.740926 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc"] Nov 25 17:02:40 crc kubenswrapper[4802]: E1125 17:02:40.741801 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="376dff9f-509e-46c3-b110-8bf2c43dffda" containerName="pull" Nov 25 17:02:40 crc kubenswrapper[4802]: I1125 17:02:40.741819 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="376dff9f-509e-46c3-b110-8bf2c43dffda" containerName="pull" Nov 25 17:02:40 crc kubenswrapper[4802]: E1125 17:02:40.741842 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="376dff9f-509e-46c3-b110-8bf2c43dffda" containerName="extract" Nov 25 17:02:40 crc kubenswrapper[4802]: I1125 17:02:40.741851 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="376dff9f-509e-46c3-b110-8bf2c43dffda" containerName="extract" Nov 25 17:02:40 crc kubenswrapper[4802]: E1125 17:02:40.741863 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="376dff9f-509e-46c3-b110-8bf2c43dffda" containerName="util" Nov 25 17:02:40 crc kubenswrapper[4802]: I1125 17:02:40.741872 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="376dff9f-509e-46c3-b110-8bf2c43dffda" containerName="util" Nov 25 17:02:40 crc kubenswrapper[4802]: I1125 17:02:40.742003 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="376dff9f-509e-46c3-b110-8bf2c43dffda" containerName="extract" Nov 25 17:02:40 crc kubenswrapper[4802]: I1125 17:02:40.742627 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" Nov 25 17:02:40 crc kubenswrapper[4802]: I1125 17:02:40.745323 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-xrjdp" Nov 25 17:02:40 crc kubenswrapper[4802]: I1125 17:02:40.759762 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc"] Nov 25 17:02:40 crc kubenswrapper[4802]: I1125 17:02:40.828316 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t4mf\" (UniqueName: \"kubernetes.io/projected/373fd8b2-c469-46b8-b20e-8ecd875c4a39-kube-api-access-5t4mf\") pod \"rabbitmq-cluster-operator-779fc9694b-987nc\" (UID: \"373fd8b2-c469-46b8-b20e-8ecd875c4a39\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" Nov 25 17:02:40 crc kubenswrapper[4802]: I1125 17:02:40.929377 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t4mf\" (UniqueName: \"kubernetes.io/projected/373fd8b2-c469-46b8-b20e-8ecd875c4a39-kube-api-access-5t4mf\") pod \"rabbitmq-cluster-operator-779fc9694b-987nc\" (UID: \"373fd8b2-c469-46b8-b20e-8ecd875c4a39\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" Nov 25 17:02:40 crc kubenswrapper[4802]: I1125 17:02:40.949144 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t4mf\" (UniqueName: \"kubernetes.io/projected/373fd8b2-c469-46b8-b20e-8ecd875c4a39-kube-api-access-5t4mf\") pod \"rabbitmq-cluster-operator-779fc9694b-987nc\" (UID: \"373fd8b2-c469-46b8-b20e-8ecd875c4a39\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" Nov 25 17:02:41 crc kubenswrapper[4802]: I1125 17:02:41.065329 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" Nov 25 17:02:41 crc kubenswrapper[4802]: I1125 17:02:41.461517 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc"] Nov 25 17:02:41 crc kubenswrapper[4802]: W1125 17:02:41.468354 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod373fd8b2_c469_46b8_b20e_8ecd875c4a39.slice/crio-ddc8a3c2576d9808a938abb794088dbe2d69c0714ded74edaeb8143b4b3d07f0 WatchSource:0}: Error finding container ddc8a3c2576d9808a938abb794088dbe2d69c0714ded74edaeb8143b4b3d07f0: Status 404 returned error can't find the container with id ddc8a3c2576d9808a938abb794088dbe2d69c0714ded74edaeb8143b4b3d07f0 Nov 25 17:02:41 crc kubenswrapper[4802]: I1125 17:02:41.858537 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" event={"ID":"373fd8b2-c469-46b8-b20e-8ecd875c4a39","Type":"ContainerStarted","Data":"ddc8a3c2576d9808a938abb794088dbe2d69c0714ded74edaeb8143b4b3d07f0"} Nov 25 17:02:44 crc kubenswrapper[4802]: I1125 17:02:44.878398 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" event={"ID":"373fd8b2-c469-46b8-b20e-8ecd875c4a39","Type":"ContainerStarted","Data":"7e4484cf13bf5fb149873ab54dc09699f84f1df1654eb8e4997afa2fb919a339"} Nov 25 17:02:44 crc kubenswrapper[4802]: I1125 17:02:44.893145 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" podStartSLOduration=1.921740154 podStartE2EDuration="4.893112996s" podCreationTimestamp="2025-11-25 17:02:40 +0000 UTC" firstStartedPulling="2025-11-25 17:02:41.472392658 +0000 UTC m=+944.616739854" lastFinishedPulling="2025-11-25 17:02:44.44376552 +0000 UTC m=+947.588112696" observedRunningTime="2025-11-25 17:02:44.892243893 +0000 UTC m=+948.036591089" watchObservedRunningTime="2025-11-25 17:02:44.893112996 +0000 UTC m=+948.037460182" Nov 25 17:02:50 crc kubenswrapper[4802]: I1125 17:02:50.950965 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 25 17:02:50 crc kubenswrapper[4802]: I1125 17:02:50.952432 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:50 crc kubenswrapper[4802]: I1125 17:02:50.954326 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Nov 25 17:02:50 crc kubenswrapper[4802]: I1125 17:02:50.954416 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Nov 25 17:02:50 crc kubenswrapper[4802]: I1125 17:02:50.954366 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Nov 25 17:02:50 crc kubenswrapper[4802]: I1125 17:02:50.954598 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Nov 25 17:02:50 crc kubenswrapper[4802]: I1125 17:02:50.955458 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-qmlbh" Nov 25 17:02:50 crc kubenswrapper[4802]: I1125 17:02:50.961690 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.068473 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74e76589-71a7-4c7f-bbac-edee84556bd7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.068533 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74e76589-71a7-4c7f-bbac-edee84556bd7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.068568 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-32c1f75f-a2b7-43b5-ab2e-4cf2dabf7877\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32c1f75f-a2b7-43b5-ab2e-4cf2dabf7877\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.068730 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74e76589-71a7-4c7f-bbac-edee84556bd7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.068771 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74e76589-71a7-4c7f-bbac-edee84556bd7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.068801 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74e76589-71a7-4c7f-bbac-edee84556bd7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.068933 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74e76589-71a7-4c7f-bbac-edee84556bd7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.068962 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmpqp\" (UniqueName: \"kubernetes.io/projected/74e76589-71a7-4c7f-bbac-edee84556bd7-kube-api-access-nmpqp\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.170768 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-32c1f75f-a2b7-43b5-ab2e-4cf2dabf7877\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32c1f75f-a2b7-43b5-ab2e-4cf2dabf7877\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.170869 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74e76589-71a7-4c7f-bbac-edee84556bd7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.170892 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74e76589-71a7-4c7f-bbac-edee84556bd7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.170917 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74e76589-71a7-4c7f-bbac-edee84556bd7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.170966 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74e76589-71a7-4c7f-bbac-edee84556bd7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.170986 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmpqp\" (UniqueName: \"kubernetes.io/projected/74e76589-71a7-4c7f-bbac-edee84556bd7-kube-api-access-nmpqp\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.171017 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74e76589-71a7-4c7f-bbac-edee84556bd7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.171045 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74e76589-71a7-4c7f-bbac-edee84556bd7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.171796 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74e76589-71a7-4c7f-bbac-edee84556bd7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.172004 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74e76589-71a7-4c7f-bbac-edee84556bd7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.172041 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74e76589-71a7-4c7f-bbac-edee84556bd7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.175635 4802 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.175678 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-32c1f75f-a2b7-43b5-ab2e-4cf2dabf7877\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32c1f75f-a2b7-43b5-ab2e-4cf2dabf7877\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f95c695b43763da7ddcf1811999e48ebac9ea62a9619a16777d02d80a87ecb52/globalmount\"" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.177656 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74e76589-71a7-4c7f-bbac-edee84556bd7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.184809 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74e76589-71a7-4c7f-bbac-edee84556bd7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.188734 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74e76589-71a7-4c7f-bbac-edee84556bd7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.209019 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmpqp\" (UniqueName: \"kubernetes.io/projected/74e76589-71a7-4c7f-bbac-edee84556bd7-kube-api-access-nmpqp\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.212298 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-32c1f75f-a2b7-43b5-ab2e-4cf2dabf7877\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32c1f75f-a2b7-43b5-ab2e-4cf2dabf7877\") pod \"rabbitmq-server-0\" (UID: \"74e76589-71a7-4c7f-bbac-edee84556bd7\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.274313 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.670380 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 25 17:02:51 crc kubenswrapper[4802]: I1125 17:02:51.926445 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"74e76589-71a7-4c7f-bbac-edee84556bd7","Type":"ContainerStarted","Data":"84b899916822340109e08266e09aeb12235c87b33825c91efc4d17421d4c2a84"} Nov 25 17:02:52 crc kubenswrapper[4802]: I1125 17:02:52.691687 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-bs696"] Nov 25 17:02:52 crc kubenswrapper[4802]: I1125 17:02:52.692877 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-bs696" Nov 25 17:02:52 crc kubenswrapper[4802]: I1125 17:02:52.695364 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-xcg2k" Nov 25 17:02:52 crc kubenswrapper[4802]: I1125 17:02:52.700614 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-bs696"] Nov 25 17:02:52 crc kubenswrapper[4802]: I1125 17:02:52.793407 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8hqc\" (UniqueName: \"kubernetes.io/projected/e9c24dd4-6edb-4e3f-90fd-5a1230fcf452-kube-api-access-r8hqc\") pod \"keystone-operator-index-bs696\" (UID: \"e9c24dd4-6edb-4e3f-90fd-5a1230fcf452\") " pod="openstack-operators/keystone-operator-index-bs696" Nov 25 17:02:52 crc kubenswrapper[4802]: I1125 17:02:52.894267 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8hqc\" (UniqueName: \"kubernetes.io/projected/e9c24dd4-6edb-4e3f-90fd-5a1230fcf452-kube-api-access-r8hqc\") pod \"keystone-operator-index-bs696\" (UID: \"e9c24dd4-6edb-4e3f-90fd-5a1230fcf452\") " pod="openstack-operators/keystone-operator-index-bs696" Nov 25 17:02:52 crc kubenswrapper[4802]: I1125 17:02:52.924240 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8hqc\" (UniqueName: \"kubernetes.io/projected/e9c24dd4-6edb-4e3f-90fd-5a1230fcf452-kube-api-access-r8hqc\") pod \"keystone-operator-index-bs696\" (UID: \"e9c24dd4-6edb-4e3f-90fd-5a1230fcf452\") " pod="openstack-operators/keystone-operator-index-bs696" Nov 25 17:02:53 crc kubenswrapper[4802]: I1125 17:02:53.011418 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-bs696" Nov 25 17:02:53 crc kubenswrapper[4802]: I1125 17:02:53.213439 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-bs696"] Nov 25 17:02:53 crc kubenswrapper[4802]: I1125 17:02:53.941729 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-bs696" event={"ID":"e9c24dd4-6edb-4e3f-90fd-5a1230fcf452","Type":"ContainerStarted","Data":"60a6f299ec4a27ae088422c28d52c94e4ab118bde215d71ab57966ef676c8c70"} Nov 25 17:02:54 crc kubenswrapper[4802]: I1125 17:02:54.248418 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 17:02:54 crc kubenswrapper[4802]: I1125 17:02:54.248782 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 17:02:58 crc kubenswrapper[4802]: I1125 17:02:58.973607 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-bs696" event={"ID":"e9c24dd4-6edb-4e3f-90fd-5a1230fcf452","Type":"ContainerStarted","Data":"c97c49f5e03735d2f93a7734c06c0854c7edff79cdc1e1b0d79432b10554bc10"} Nov 25 17:02:58 crc kubenswrapper[4802]: I1125 17:02:58.993566 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-bs696" podStartSLOduration=2.628918608 podStartE2EDuration="6.993545825s" podCreationTimestamp="2025-11-25 17:02:52 +0000 UTC" firstStartedPulling="2025-11-25 17:02:53.225887961 +0000 UTC m=+956.370235147" lastFinishedPulling="2025-11-25 17:02:57.590515178 +0000 UTC m=+960.734862364" observedRunningTime="2025-11-25 17:02:58.989313311 +0000 UTC m=+962.133660537" watchObservedRunningTime="2025-11-25 17:02:58.993545825 +0000 UTC m=+962.137893011" Nov 25 17:03:00 crc kubenswrapper[4802]: I1125 17:03:00.986624 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"74e76589-71a7-4c7f-bbac-edee84556bd7","Type":"ContainerStarted","Data":"19c18a6c2b389681c03472b1a85bd59301f09a9d070d6a7c51dace9202981b01"} Nov 25 17:03:03 crc kubenswrapper[4802]: I1125 17:03:03.012314 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-bs696" Nov 25 17:03:03 crc kubenswrapper[4802]: I1125 17:03:03.012655 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-bs696" Nov 25 17:03:03 crc kubenswrapper[4802]: I1125 17:03:03.034011 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-bs696" Nov 25 17:03:04 crc kubenswrapper[4802]: I1125 17:03:04.026963 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-bs696" Nov 25 17:03:10 crc kubenswrapper[4802]: I1125 17:03:10.735941 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l"] Nov 25 17:03:10 crc kubenswrapper[4802]: I1125 17:03:10.737520 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l" Nov 25 17:03:10 crc kubenswrapper[4802]: I1125 17:03:10.739592 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-t7662" Nov 25 17:03:10 crc kubenswrapper[4802]: I1125 17:03:10.746477 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l"] Nov 25 17:03:10 crc kubenswrapper[4802]: I1125 17:03:10.866048 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp4t6\" (UniqueName: \"kubernetes.io/projected/2e812eef-efff-4d2b-a943-57b1305d73eb-kube-api-access-bp4t6\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l\" (UID: \"2e812eef-efff-4d2b-a943-57b1305d73eb\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l" Nov 25 17:03:10 crc kubenswrapper[4802]: I1125 17:03:10.866422 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2e812eef-efff-4d2b-a943-57b1305d73eb-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l\" (UID: \"2e812eef-efff-4d2b-a943-57b1305d73eb\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l" Nov 25 17:03:10 crc kubenswrapper[4802]: I1125 17:03:10.866548 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2e812eef-efff-4d2b-a943-57b1305d73eb-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l\" (UID: \"2e812eef-efff-4d2b-a943-57b1305d73eb\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l" Nov 25 17:03:10 crc kubenswrapper[4802]: I1125 17:03:10.967956 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp4t6\" (UniqueName: \"kubernetes.io/projected/2e812eef-efff-4d2b-a943-57b1305d73eb-kube-api-access-bp4t6\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l\" (UID: \"2e812eef-efff-4d2b-a943-57b1305d73eb\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l" Nov 25 17:03:10 crc kubenswrapper[4802]: I1125 17:03:10.968361 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2e812eef-efff-4d2b-a943-57b1305d73eb-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l\" (UID: \"2e812eef-efff-4d2b-a943-57b1305d73eb\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l" Nov 25 17:03:10 crc kubenswrapper[4802]: I1125 17:03:10.968481 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2e812eef-efff-4d2b-a943-57b1305d73eb-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l\" (UID: \"2e812eef-efff-4d2b-a943-57b1305d73eb\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l" Nov 25 17:03:10 crc kubenswrapper[4802]: I1125 17:03:10.968878 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2e812eef-efff-4d2b-a943-57b1305d73eb-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l\" (UID: \"2e812eef-efff-4d2b-a943-57b1305d73eb\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l" Nov 25 17:03:10 crc kubenswrapper[4802]: I1125 17:03:10.969016 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2e812eef-efff-4d2b-a943-57b1305d73eb-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l\" (UID: \"2e812eef-efff-4d2b-a943-57b1305d73eb\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l" Nov 25 17:03:10 crc kubenswrapper[4802]: I1125 17:03:10.990158 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp4t6\" (UniqueName: \"kubernetes.io/projected/2e812eef-efff-4d2b-a943-57b1305d73eb-kube-api-access-bp4t6\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l\" (UID: \"2e812eef-efff-4d2b-a943-57b1305d73eb\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l" Nov 25 17:03:11 crc kubenswrapper[4802]: I1125 17:03:11.053933 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l" Nov 25 17:03:11 crc kubenswrapper[4802]: W1125 17:03:11.519209 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e812eef_efff_4d2b_a943_57b1305d73eb.slice/crio-c3e40bfc20894f4bee1d1a6b8cac421fe79b1a4de1ada397f988117020fba115 WatchSource:0}: Error finding container c3e40bfc20894f4bee1d1a6b8cac421fe79b1a4de1ada397f988117020fba115: Status 404 returned error can't find the container with id c3e40bfc20894f4bee1d1a6b8cac421fe79b1a4de1ada397f988117020fba115 Nov 25 17:03:11 crc kubenswrapper[4802]: I1125 17:03:11.520558 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l"] Nov 25 17:03:12 crc kubenswrapper[4802]: I1125 17:03:12.058387 4802 generic.go:334] "Generic (PLEG): container finished" podID="2e812eef-efff-4d2b-a943-57b1305d73eb" containerID="69de3778145cd572644cae1bf910cc400c7a9ae29fff91fd8037304439223c68" exitCode=0 Nov 25 17:03:12 crc kubenswrapper[4802]: I1125 17:03:12.058454 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l" event={"ID":"2e812eef-efff-4d2b-a943-57b1305d73eb","Type":"ContainerDied","Data":"69de3778145cd572644cae1bf910cc400c7a9ae29fff91fd8037304439223c68"} Nov 25 17:03:12 crc kubenswrapper[4802]: I1125 17:03:12.058683 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l" event={"ID":"2e812eef-efff-4d2b-a943-57b1305d73eb","Type":"ContainerStarted","Data":"c3e40bfc20894f4bee1d1a6b8cac421fe79b1a4de1ada397f988117020fba115"} Nov 25 17:03:14 crc kubenswrapper[4802]: I1125 17:03:14.073944 4802 generic.go:334] "Generic (PLEG): container finished" podID="2e812eef-efff-4d2b-a943-57b1305d73eb" containerID="2ba4d80ff55b3a1d5fa75380438b66b7314666f65f260e753b3089e5061111cb" exitCode=0 Nov 25 17:03:14 crc kubenswrapper[4802]: I1125 17:03:14.074025 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l" event={"ID":"2e812eef-efff-4d2b-a943-57b1305d73eb","Type":"ContainerDied","Data":"2ba4d80ff55b3a1d5fa75380438b66b7314666f65f260e753b3089e5061111cb"} Nov 25 17:03:15 crc kubenswrapper[4802]: I1125 17:03:15.087420 4802 generic.go:334] "Generic (PLEG): container finished" podID="2e812eef-efff-4d2b-a943-57b1305d73eb" containerID="0979ded7c6b8110cea8c915460b3f1fc6e01d04686cef74e9a717abebbc4c8b7" exitCode=0 Nov 25 17:03:15 crc kubenswrapper[4802]: I1125 17:03:15.087490 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l" event={"ID":"2e812eef-efff-4d2b-a943-57b1305d73eb","Type":"ContainerDied","Data":"0979ded7c6b8110cea8c915460b3f1fc6e01d04686cef74e9a717abebbc4c8b7"} Nov 25 17:03:16 crc kubenswrapper[4802]: I1125 17:03:16.396757 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l" Nov 25 17:03:16 crc kubenswrapper[4802]: I1125 17:03:16.545238 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2e812eef-efff-4d2b-a943-57b1305d73eb-util\") pod \"2e812eef-efff-4d2b-a943-57b1305d73eb\" (UID: \"2e812eef-efff-4d2b-a943-57b1305d73eb\") " Nov 25 17:03:16 crc kubenswrapper[4802]: I1125 17:03:16.545318 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bp4t6\" (UniqueName: \"kubernetes.io/projected/2e812eef-efff-4d2b-a943-57b1305d73eb-kube-api-access-bp4t6\") pod \"2e812eef-efff-4d2b-a943-57b1305d73eb\" (UID: \"2e812eef-efff-4d2b-a943-57b1305d73eb\") " Nov 25 17:03:16 crc kubenswrapper[4802]: I1125 17:03:16.545410 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2e812eef-efff-4d2b-a943-57b1305d73eb-bundle\") pod \"2e812eef-efff-4d2b-a943-57b1305d73eb\" (UID: \"2e812eef-efff-4d2b-a943-57b1305d73eb\") " Nov 25 17:03:16 crc kubenswrapper[4802]: I1125 17:03:16.546364 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e812eef-efff-4d2b-a943-57b1305d73eb-bundle" (OuterVolumeSpecName: "bundle") pod "2e812eef-efff-4d2b-a943-57b1305d73eb" (UID: "2e812eef-efff-4d2b-a943-57b1305d73eb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:03:16 crc kubenswrapper[4802]: I1125 17:03:16.553987 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e812eef-efff-4d2b-a943-57b1305d73eb-kube-api-access-bp4t6" (OuterVolumeSpecName: "kube-api-access-bp4t6") pod "2e812eef-efff-4d2b-a943-57b1305d73eb" (UID: "2e812eef-efff-4d2b-a943-57b1305d73eb"). InnerVolumeSpecName "kube-api-access-bp4t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:03:16 crc kubenswrapper[4802]: I1125 17:03:16.558696 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e812eef-efff-4d2b-a943-57b1305d73eb-util" (OuterVolumeSpecName: "util") pod "2e812eef-efff-4d2b-a943-57b1305d73eb" (UID: "2e812eef-efff-4d2b-a943-57b1305d73eb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:03:16 crc kubenswrapper[4802]: I1125 17:03:16.647693 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bp4t6\" (UniqueName: \"kubernetes.io/projected/2e812eef-efff-4d2b-a943-57b1305d73eb-kube-api-access-bp4t6\") on node \"crc\" DevicePath \"\"" Nov 25 17:03:16 crc kubenswrapper[4802]: I1125 17:03:16.647738 4802 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2e812eef-efff-4d2b-a943-57b1305d73eb-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 17:03:16 crc kubenswrapper[4802]: I1125 17:03:16.647762 4802 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2e812eef-efff-4d2b-a943-57b1305d73eb-util\") on node \"crc\" DevicePath \"\"" Nov 25 17:03:17 crc kubenswrapper[4802]: I1125 17:03:17.103065 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l" event={"ID":"2e812eef-efff-4d2b-a943-57b1305d73eb","Type":"ContainerDied","Data":"c3e40bfc20894f4bee1d1a6b8cac421fe79b1a4de1ada397f988117020fba115"} Nov 25 17:03:17 crc kubenswrapper[4802]: I1125 17:03:17.103161 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3e40bfc20894f4bee1d1a6b8cac421fe79b1a4de1ada397f988117020fba115" Nov 25 17:03:17 crc kubenswrapper[4802]: I1125 17:03:17.103262 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l" Nov 25 17:03:24 crc kubenswrapper[4802]: I1125 17:03:24.249320 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 17:03:24 crc kubenswrapper[4802]: I1125 17:03:24.250252 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 17:03:24 crc kubenswrapper[4802]: I1125 17:03:24.250323 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 17:03:24 crc kubenswrapper[4802]: I1125 17:03:24.251200 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b903688e55473d1e5bdabb3f4b23a6b1ac78c87eaabad518c885de28866fe47e"} pod="openshift-machine-config-operator/machine-config-daemon-h29wc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 17:03:24 crc kubenswrapper[4802]: I1125 17:03:24.251266 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" containerID="cri-o://b903688e55473d1e5bdabb3f4b23a6b1ac78c87eaabad518c885de28866fe47e" gracePeriod=600 Nov 25 17:03:25 crc kubenswrapper[4802]: I1125 17:03:25.299232 4802 generic.go:334] "Generic (PLEG): container finished" podID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerID="b903688e55473d1e5bdabb3f4b23a6b1ac78c87eaabad518c885de28866fe47e" exitCode=0 Nov 25 17:03:25 crc kubenswrapper[4802]: I1125 17:03:25.299267 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerDied","Data":"b903688e55473d1e5bdabb3f4b23a6b1ac78c87eaabad518c885de28866fe47e"} Nov 25 17:03:25 crc kubenswrapper[4802]: I1125 17:03:25.299774 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerStarted","Data":"18e081a4af641b516662b68a934da8c320eec5e2196f1740999f260df1353261"} Nov 25 17:03:25 crc kubenswrapper[4802]: I1125 17:03:25.299796 4802 scope.go:117] "RemoveContainer" containerID="2dd6a8bbe96b494880e307bbcf47128d8479e3413bfeccced18011d7c6e2daae" Nov 25 17:03:28 crc kubenswrapper[4802]: I1125 17:03:28.781570 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr"] Nov 25 17:03:28 crc kubenswrapper[4802]: E1125 17:03:28.782435 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e812eef-efff-4d2b-a943-57b1305d73eb" containerName="pull" Nov 25 17:03:28 crc kubenswrapper[4802]: I1125 17:03:28.782453 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e812eef-efff-4d2b-a943-57b1305d73eb" containerName="pull" Nov 25 17:03:28 crc kubenswrapper[4802]: E1125 17:03:28.782473 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e812eef-efff-4d2b-a943-57b1305d73eb" containerName="util" Nov 25 17:03:28 crc kubenswrapper[4802]: I1125 17:03:28.782481 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e812eef-efff-4d2b-a943-57b1305d73eb" containerName="util" Nov 25 17:03:28 crc kubenswrapper[4802]: E1125 17:03:28.782502 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e812eef-efff-4d2b-a943-57b1305d73eb" containerName="extract" Nov 25 17:03:28 crc kubenswrapper[4802]: I1125 17:03:28.782510 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e812eef-efff-4d2b-a943-57b1305d73eb" containerName="extract" Nov 25 17:03:28 crc kubenswrapper[4802]: I1125 17:03:28.782636 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e812eef-efff-4d2b-a943-57b1305d73eb" containerName="extract" Nov 25 17:03:28 crc kubenswrapper[4802]: I1125 17:03:28.783213 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:03:28 crc kubenswrapper[4802]: I1125 17:03:28.788416 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Nov 25 17:03:28 crc kubenswrapper[4802]: I1125 17:03:28.788423 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-s9mkz" Nov 25 17:03:28 crc kubenswrapper[4802]: I1125 17:03:28.807390 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr"] Nov 25 17:03:28 crc kubenswrapper[4802]: I1125 17:03:28.855459 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zhdj\" (UniqueName: \"kubernetes.io/projected/a5bdabd0-7540-4d45-8675-9bd88c415957-kube-api-access-2zhdj\") pod \"keystone-operator-controller-manager-65c6fdb94d-5cfbr\" (UID: \"a5bdabd0-7540-4d45-8675-9bd88c415957\") " pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:03:28 crc kubenswrapper[4802]: I1125 17:03:28.855664 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a5bdabd0-7540-4d45-8675-9bd88c415957-webhook-cert\") pod \"keystone-operator-controller-manager-65c6fdb94d-5cfbr\" (UID: \"a5bdabd0-7540-4d45-8675-9bd88c415957\") " pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:03:28 crc kubenswrapper[4802]: I1125 17:03:28.855720 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a5bdabd0-7540-4d45-8675-9bd88c415957-apiservice-cert\") pod \"keystone-operator-controller-manager-65c6fdb94d-5cfbr\" (UID: \"a5bdabd0-7540-4d45-8675-9bd88c415957\") " pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:03:28 crc kubenswrapper[4802]: I1125 17:03:28.956285 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a5bdabd0-7540-4d45-8675-9bd88c415957-webhook-cert\") pod \"keystone-operator-controller-manager-65c6fdb94d-5cfbr\" (UID: \"a5bdabd0-7540-4d45-8675-9bd88c415957\") " pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:03:28 crc kubenswrapper[4802]: I1125 17:03:28.956377 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a5bdabd0-7540-4d45-8675-9bd88c415957-apiservice-cert\") pod \"keystone-operator-controller-manager-65c6fdb94d-5cfbr\" (UID: \"a5bdabd0-7540-4d45-8675-9bd88c415957\") " pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:03:28 crc kubenswrapper[4802]: I1125 17:03:28.956428 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zhdj\" (UniqueName: \"kubernetes.io/projected/a5bdabd0-7540-4d45-8675-9bd88c415957-kube-api-access-2zhdj\") pod \"keystone-operator-controller-manager-65c6fdb94d-5cfbr\" (UID: \"a5bdabd0-7540-4d45-8675-9bd88c415957\") " pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:03:28 crc kubenswrapper[4802]: I1125 17:03:28.977290 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a5bdabd0-7540-4d45-8675-9bd88c415957-webhook-cert\") pod \"keystone-operator-controller-manager-65c6fdb94d-5cfbr\" (UID: \"a5bdabd0-7540-4d45-8675-9bd88c415957\") " pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:03:28 crc kubenswrapper[4802]: I1125 17:03:28.977352 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a5bdabd0-7540-4d45-8675-9bd88c415957-apiservice-cert\") pod \"keystone-operator-controller-manager-65c6fdb94d-5cfbr\" (UID: \"a5bdabd0-7540-4d45-8675-9bd88c415957\") " pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:03:28 crc kubenswrapper[4802]: I1125 17:03:28.982086 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zhdj\" (UniqueName: \"kubernetes.io/projected/a5bdabd0-7540-4d45-8675-9bd88c415957-kube-api-access-2zhdj\") pod \"keystone-operator-controller-manager-65c6fdb94d-5cfbr\" (UID: \"a5bdabd0-7540-4d45-8675-9bd88c415957\") " pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:03:29 crc kubenswrapper[4802]: I1125 17:03:29.105174 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:03:29 crc kubenswrapper[4802]: I1125 17:03:29.518330 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr"] Nov 25 17:03:30 crc kubenswrapper[4802]: I1125 17:03:30.348540 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" event={"ID":"a5bdabd0-7540-4d45-8675-9bd88c415957","Type":"ContainerStarted","Data":"45fc36d19b5f8ae9c1d5d08e083f812f8361593198feee3c799a6e7db1729c7c"} Nov 25 17:03:32 crc kubenswrapper[4802]: I1125 17:03:32.376865 4802 generic.go:334] "Generic (PLEG): container finished" podID="74e76589-71a7-4c7f-bbac-edee84556bd7" containerID="19c18a6c2b389681c03472b1a85bd59301f09a9d070d6a7c51dace9202981b01" exitCode=0 Nov 25 17:03:32 crc kubenswrapper[4802]: I1125 17:03:32.376953 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"74e76589-71a7-4c7f-bbac-edee84556bd7","Type":"ContainerDied","Data":"19c18a6c2b389681c03472b1a85bd59301f09a9d070d6a7c51dace9202981b01"} Nov 25 17:03:33 crc kubenswrapper[4802]: I1125 17:03:33.384337 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"74e76589-71a7-4c7f-bbac-edee84556bd7","Type":"ContainerStarted","Data":"fcb269685cf78c3a6807fcb54cde15d509a574fcfeb1f53dc9d705ec6f9b6c85"} Nov 25 17:03:33 crc kubenswrapper[4802]: I1125 17:03:33.384724 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:03:33 crc kubenswrapper[4802]: I1125 17:03:33.386902 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" event={"ID":"a5bdabd0-7540-4d45-8675-9bd88c415957","Type":"ContainerStarted","Data":"c047bba0964146b17129c401fb29cd7d6ccbedf4efe035eab267e9821dfc0d58"} Nov 25 17:03:33 crc kubenswrapper[4802]: I1125 17:03:33.387029 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:03:33 crc kubenswrapper[4802]: I1125 17:03:33.411443 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/rabbitmq-server-0" podStartSLOduration=37.01649764 podStartE2EDuration="44.41142608s" podCreationTimestamp="2025-11-25 17:02:49 +0000 UTC" firstStartedPulling="2025-11-25 17:02:51.685061605 +0000 UTC m=+954.829408791" lastFinishedPulling="2025-11-25 17:02:59.079990045 +0000 UTC m=+962.224337231" observedRunningTime="2025-11-25 17:03:33.410357812 +0000 UTC m=+996.554704998" watchObservedRunningTime="2025-11-25 17:03:33.41142608 +0000 UTC m=+996.555773266" Nov 25 17:03:33 crc kubenswrapper[4802]: I1125 17:03:33.426677 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" podStartSLOduration=2.108658208 podStartE2EDuration="5.426658043s" podCreationTimestamp="2025-11-25 17:03:28 +0000 UTC" firstStartedPulling="2025-11-25 17:03:29.525484382 +0000 UTC m=+992.669831568" lastFinishedPulling="2025-11-25 17:03:32.843484217 +0000 UTC m=+995.987831403" observedRunningTime="2025-11-25 17:03:33.426493259 +0000 UTC m=+996.570840455" watchObservedRunningTime="2025-11-25 17:03:33.426658043 +0000 UTC m=+996.571005219" Nov 25 17:03:39 crc kubenswrapper[4802]: I1125 17:03:39.109494 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.238290 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-6222-account-create-update-qzmx5"] Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.239731 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-6222-account-create-update-qzmx5" Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.241627 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-db-secret" Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.246644 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-6222-account-create-update-qzmx5"] Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.331348 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-create-gzvs6"] Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.332357 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-gzvs6" Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.341667 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-gzvs6"] Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.346081 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7kms\" (UniqueName: \"kubernetes.io/projected/9e6a7e14-e3b5-4105-8efb-25f2fd871dd6-kube-api-access-b7kms\") pod \"keystone-6222-account-create-update-qzmx5\" (UID: \"9e6a7e14-e3b5-4105-8efb-25f2fd871dd6\") " pod="glance-kuttl-tests/keystone-6222-account-create-update-qzmx5" Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.346163 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e6a7e14-e3b5-4105-8efb-25f2fd871dd6-operator-scripts\") pod \"keystone-6222-account-create-update-qzmx5\" (UID: \"9e6a7e14-e3b5-4105-8efb-25f2fd871dd6\") " pod="glance-kuttl-tests/keystone-6222-account-create-update-qzmx5" Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.447707 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2004be61-2b02-4bc4-a79e-5beba8cb4a60-operator-scripts\") pod \"keystone-db-create-gzvs6\" (UID: \"2004be61-2b02-4bc4-a79e-5beba8cb4a60\") " pod="glance-kuttl-tests/keystone-db-create-gzvs6" Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.447887 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7kms\" (UniqueName: \"kubernetes.io/projected/9e6a7e14-e3b5-4105-8efb-25f2fd871dd6-kube-api-access-b7kms\") pod \"keystone-6222-account-create-update-qzmx5\" (UID: \"9e6a7e14-e3b5-4105-8efb-25f2fd871dd6\") " pod="glance-kuttl-tests/keystone-6222-account-create-update-qzmx5" Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.447932 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cdxz\" (UniqueName: \"kubernetes.io/projected/2004be61-2b02-4bc4-a79e-5beba8cb4a60-kube-api-access-5cdxz\") pod \"keystone-db-create-gzvs6\" (UID: \"2004be61-2b02-4bc4-a79e-5beba8cb4a60\") " pod="glance-kuttl-tests/keystone-db-create-gzvs6" Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.447990 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e6a7e14-e3b5-4105-8efb-25f2fd871dd6-operator-scripts\") pod \"keystone-6222-account-create-update-qzmx5\" (UID: \"9e6a7e14-e3b5-4105-8efb-25f2fd871dd6\") " pod="glance-kuttl-tests/keystone-6222-account-create-update-qzmx5" Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.448841 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e6a7e14-e3b5-4105-8efb-25f2fd871dd6-operator-scripts\") pod \"keystone-6222-account-create-update-qzmx5\" (UID: \"9e6a7e14-e3b5-4105-8efb-25f2fd871dd6\") " pod="glance-kuttl-tests/keystone-6222-account-create-update-qzmx5" Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.465143 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7kms\" (UniqueName: \"kubernetes.io/projected/9e6a7e14-e3b5-4105-8efb-25f2fd871dd6-kube-api-access-b7kms\") pod \"keystone-6222-account-create-update-qzmx5\" (UID: \"9e6a7e14-e3b5-4105-8efb-25f2fd871dd6\") " pod="glance-kuttl-tests/keystone-6222-account-create-update-qzmx5" Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.549666 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2004be61-2b02-4bc4-a79e-5beba8cb4a60-operator-scripts\") pod \"keystone-db-create-gzvs6\" (UID: \"2004be61-2b02-4bc4-a79e-5beba8cb4a60\") " pod="glance-kuttl-tests/keystone-db-create-gzvs6" Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.549792 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cdxz\" (UniqueName: \"kubernetes.io/projected/2004be61-2b02-4bc4-a79e-5beba8cb4a60-kube-api-access-5cdxz\") pod \"keystone-db-create-gzvs6\" (UID: \"2004be61-2b02-4bc4-a79e-5beba8cb4a60\") " pod="glance-kuttl-tests/keystone-db-create-gzvs6" Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.550851 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2004be61-2b02-4bc4-a79e-5beba8cb4a60-operator-scripts\") pod \"keystone-db-create-gzvs6\" (UID: \"2004be61-2b02-4bc4-a79e-5beba8cb4a60\") " pod="glance-kuttl-tests/keystone-db-create-gzvs6" Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.564154 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-6222-account-create-update-qzmx5" Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.566207 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cdxz\" (UniqueName: \"kubernetes.io/projected/2004be61-2b02-4bc4-a79e-5beba8cb4a60-kube-api-access-5cdxz\") pod \"keystone-db-create-gzvs6\" (UID: \"2004be61-2b02-4bc4-a79e-5beba8cb4a60\") " pod="glance-kuttl-tests/keystone-db-create-gzvs6" Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.648898 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-gzvs6" Nov 25 17:03:42 crc kubenswrapper[4802]: I1125 17:03:42.927302 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-gzvs6"] Nov 25 17:03:42 crc kubenswrapper[4802]: W1125 17:03:42.932814 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2004be61_2b02_4bc4_a79e_5beba8cb4a60.slice/crio-7cd1bb40a7fecc9e88541bc17ba1770a8f1bca91faaa599ec428386fdce0ad67 WatchSource:0}: Error finding container 7cd1bb40a7fecc9e88541bc17ba1770a8f1bca91faaa599ec428386fdce0ad67: Status 404 returned error can't find the container with id 7cd1bb40a7fecc9e88541bc17ba1770a8f1bca91faaa599ec428386fdce0ad67 Nov 25 17:03:43 crc kubenswrapper[4802]: I1125 17:03:43.080025 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-6222-account-create-update-qzmx5"] Nov 25 17:03:43 crc kubenswrapper[4802]: W1125 17:03:43.085042 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e6a7e14_e3b5_4105_8efb_25f2fd871dd6.slice/crio-9cf2221b6a59eea93f47c27d7b1e70df90e8a6a9475f073a92a2a526658a5049 WatchSource:0}: Error finding container 9cf2221b6a59eea93f47c27d7b1e70df90e8a6a9475f073a92a2a526658a5049: Status 404 returned error can't find the container with id 9cf2221b6a59eea93f47c27d7b1e70df90e8a6a9475f073a92a2a526658a5049 Nov 25 17:03:43 crc kubenswrapper[4802]: I1125 17:03:43.447086 4802 generic.go:334] "Generic (PLEG): container finished" podID="2004be61-2b02-4bc4-a79e-5beba8cb4a60" containerID="e3c7904150153962589b5788cab2aa5b01c3f28c73938dde9f6e0b9c94461b54" exitCode=0 Nov 25 17:03:43 crc kubenswrapper[4802]: I1125 17:03:43.447189 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-gzvs6" event={"ID":"2004be61-2b02-4bc4-a79e-5beba8cb4a60","Type":"ContainerDied","Data":"e3c7904150153962589b5788cab2aa5b01c3f28c73938dde9f6e0b9c94461b54"} Nov 25 17:03:43 crc kubenswrapper[4802]: I1125 17:03:43.447596 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-gzvs6" event={"ID":"2004be61-2b02-4bc4-a79e-5beba8cb4a60","Type":"ContainerStarted","Data":"7cd1bb40a7fecc9e88541bc17ba1770a8f1bca91faaa599ec428386fdce0ad67"} Nov 25 17:03:43 crc kubenswrapper[4802]: I1125 17:03:43.449355 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-6222-account-create-update-qzmx5" event={"ID":"9e6a7e14-e3b5-4105-8efb-25f2fd871dd6","Type":"ContainerStarted","Data":"6d230defa21cecb2df57a1df15e4866fc3644483b542b239c2a0dfbb84e2bff3"} Nov 25 17:03:43 crc kubenswrapper[4802]: I1125 17:03:43.449410 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-6222-account-create-update-qzmx5" event={"ID":"9e6a7e14-e3b5-4105-8efb-25f2fd871dd6","Type":"ContainerStarted","Data":"9cf2221b6a59eea93f47c27d7b1e70df90e8a6a9475f073a92a2a526658a5049"} Nov 25 17:03:43 crc kubenswrapper[4802]: I1125 17:03:43.480359 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-6222-account-create-update-qzmx5" podStartSLOduration=1.480333657 podStartE2EDuration="1.480333657s" podCreationTimestamp="2025-11-25 17:03:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:03:43.474675244 +0000 UTC m=+1006.619022430" watchObservedRunningTime="2025-11-25 17:03:43.480333657 +0000 UTC m=+1006.624680843" Nov 25 17:03:43 crc kubenswrapper[4802]: I1125 17:03:43.693761 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-xrnj5"] Nov 25 17:03:43 crc kubenswrapper[4802]: I1125 17:03:43.694762 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-xrnj5" Nov 25 17:03:43 crc kubenswrapper[4802]: I1125 17:03:43.697292 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-xh9p6" Nov 25 17:03:43 crc kubenswrapper[4802]: I1125 17:03:43.700790 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-xrnj5"] Nov 25 17:03:43 crc kubenswrapper[4802]: I1125 17:03:43.869395 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tq9c\" (UniqueName: \"kubernetes.io/projected/750a14b5-f70f-4945-9e4a-09e4e527ae62-kube-api-access-5tq9c\") pod \"horizon-operator-index-xrnj5\" (UID: \"750a14b5-f70f-4945-9e4a-09e4e527ae62\") " pod="openstack-operators/horizon-operator-index-xrnj5" Nov 25 17:03:43 crc kubenswrapper[4802]: I1125 17:03:43.971224 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tq9c\" (UniqueName: \"kubernetes.io/projected/750a14b5-f70f-4945-9e4a-09e4e527ae62-kube-api-access-5tq9c\") pod \"horizon-operator-index-xrnj5\" (UID: \"750a14b5-f70f-4945-9e4a-09e4e527ae62\") " pod="openstack-operators/horizon-operator-index-xrnj5" Nov 25 17:03:43 crc kubenswrapper[4802]: I1125 17:03:43.991460 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tq9c\" (UniqueName: \"kubernetes.io/projected/750a14b5-f70f-4945-9e4a-09e4e527ae62-kube-api-access-5tq9c\") pod \"horizon-operator-index-xrnj5\" (UID: \"750a14b5-f70f-4945-9e4a-09e4e527ae62\") " pod="openstack-operators/horizon-operator-index-xrnj5" Nov 25 17:03:44 crc kubenswrapper[4802]: I1125 17:03:44.011674 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-xrnj5" Nov 25 17:03:44 crc kubenswrapper[4802]: I1125 17:03:44.218776 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-xrnj5"] Nov 25 17:03:44 crc kubenswrapper[4802]: I1125 17:03:44.462286 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-xrnj5" event={"ID":"750a14b5-f70f-4945-9e4a-09e4e527ae62","Type":"ContainerStarted","Data":"c5f2969d40e26a177ba2f0fc0e4c1bf75c74004050a4a9508db343755bc5424c"} Nov 25 17:03:44 crc kubenswrapper[4802]: I1125 17:03:44.464750 4802 generic.go:334] "Generic (PLEG): container finished" podID="9e6a7e14-e3b5-4105-8efb-25f2fd871dd6" containerID="6d230defa21cecb2df57a1df15e4866fc3644483b542b239c2a0dfbb84e2bff3" exitCode=0 Nov 25 17:03:44 crc kubenswrapper[4802]: I1125 17:03:44.464844 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-6222-account-create-update-qzmx5" event={"ID":"9e6a7e14-e3b5-4105-8efb-25f2fd871dd6","Type":"ContainerDied","Data":"6d230defa21cecb2df57a1df15e4866fc3644483b542b239c2a0dfbb84e2bff3"} Nov 25 17:03:44 crc kubenswrapper[4802]: I1125 17:03:44.718402 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-gzvs6" Nov 25 17:03:44 crc kubenswrapper[4802]: I1125 17:03:44.884481 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2004be61-2b02-4bc4-a79e-5beba8cb4a60-operator-scripts\") pod \"2004be61-2b02-4bc4-a79e-5beba8cb4a60\" (UID: \"2004be61-2b02-4bc4-a79e-5beba8cb4a60\") " Nov 25 17:03:44 crc kubenswrapper[4802]: I1125 17:03:44.884907 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cdxz\" (UniqueName: \"kubernetes.io/projected/2004be61-2b02-4bc4-a79e-5beba8cb4a60-kube-api-access-5cdxz\") pod \"2004be61-2b02-4bc4-a79e-5beba8cb4a60\" (UID: \"2004be61-2b02-4bc4-a79e-5beba8cb4a60\") " Nov 25 17:03:44 crc kubenswrapper[4802]: I1125 17:03:44.885526 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2004be61-2b02-4bc4-a79e-5beba8cb4a60-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2004be61-2b02-4bc4-a79e-5beba8cb4a60" (UID: "2004be61-2b02-4bc4-a79e-5beba8cb4a60"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:03:44 crc kubenswrapper[4802]: I1125 17:03:44.896283 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2004be61-2b02-4bc4-a79e-5beba8cb4a60-kube-api-access-5cdxz" (OuterVolumeSpecName: "kube-api-access-5cdxz") pod "2004be61-2b02-4bc4-a79e-5beba8cb4a60" (UID: "2004be61-2b02-4bc4-a79e-5beba8cb4a60"). InnerVolumeSpecName "kube-api-access-5cdxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:03:44 crc kubenswrapper[4802]: I1125 17:03:44.986791 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2004be61-2b02-4bc4-a79e-5beba8cb4a60-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:03:44 crc kubenswrapper[4802]: I1125 17:03:44.986829 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cdxz\" (UniqueName: \"kubernetes.io/projected/2004be61-2b02-4bc4-a79e-5beba8cb4a60-kube-api-access-5cdxz\") on node \"crc\" DevicePath \"\"" Nov 25 17:03:45 crc kubenswrapper[4802]: I1125 17:03:45.475699 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-xrnj5" event={"ID":"750a14b5-f70f-4945-9e4a-09e4e527ae62","Type":"ContainerStarted","Data":"a1cd5e9ab4ba82e1659afe7f8372e313322719dae0997cee11960b4bc99bb5b3"} Nov 25 17:03:45 crc kubenswrapper[4802]: I1125 17:03:45.479150 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-gzvs6" event={"ID":"2004be61-2b02-4bc4-a79e-5beba8cb4a60","Type":"ContainerDied","Data":"7cd1bb40a7fecc9e88541bc17ba1770a8f1bca91faaa599ec428386fdce0ad67"} Nov 25 17:03:45 crc kubenswrapper[4802]: I1125 17:03:45.479188 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-gzvs6" Nov 25 17:03:45 crc kubenswrapper[4802]: I1125 17:03:45.479205 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cd1bb40a7fecc9e88541bc17ba1770a8f1bca91faaa599ec428386fdce0ad67" Nov 25 17:03:45 crc kubenswrapper[4802]: I1125 17:03:45.504533 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-xrnj5" podStartSLOduration=1.6199219839999999 podStartE2EDuration="2.504512335s" podCreationTimestamp="2025-11-25 17:03:43 +0000 UTC" firstStartedPulling="2025-11-25 17:03:44.223080685 +0000 UTC m=+1007.367427861" lastFinishedPulling="2025-11-25 17:03:45.107671026 +0000 UTC m=+1008.252018212" observedRunningTime="2025-11-25 17:03:45.500806106 +0000 UTC m=+1008.645153292" watchObservedRunningTime="2025-11-25 17:03:45.504512335 +0000 UTC m=+1008.648859521" Nov 25 17:03:45 crc kubenswrapper[4802]: I1125 17:03:45.775696 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-6222-account-create-update-qzmx5" Nov 25 17:03:45 crc kubenswrapper[4802]: I1125 17:03:45.902001 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7kms\" (UniqueName: \"kubernetes.io/projected/9e6a7e14-e3b5-4105-8efb-25f2fd871dd6-kube-api-access-b7kms\") pod \"9e6a7e14-e3b5-4105-8efb-25f2fd871dd6\" (UID: \"9e6a7e14-e3b5-4105-8efb-25f2fd871dd6\") " Nov 25 17:03:45 crc kubenswrapper[4802]: I1125 17:03:45.902424 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e6a7e14-e3b5-4105-8efb-25f2fd871dd6-operator-scripts\") pod \"9e6a7e14-e3b5-4105-8efb-25f2fd871dd6\" (UID: \"9e6a7e14-e3b5-4105-8efb-25f2fd871dd6\") " Nov 25 17:03:45 crc kubenswrapper[4802]: I1125 17:03:45.902996 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e6a7e14-e3b5-4105-8efb-25f2fd871dd6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9e6a7e14-e3b5-4105-8efb-25f2fd871dd6" (UID: "9e6a7e14-e3b5-4105-8efb-25f2fd871dd6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:03:45 crc kubenswrapper[4802]: I1125 17:03:45.907221 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e6a7e14-e3b5-4105-8efb-25f2fd871dd6-kube-api-access-b7kms" (OuterVolumeSpecName: "kube-api-access-b7kms") pod "9e6a7e14-e3b5-4105-8efb-25f2fd871dd6" (UID: "9e6a7e14-e3b5-4105-8efb-25f2fd871dd6"). InnerVolumeSpecName "kube-api-access-b7kms". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:03:46 crc kubenswrapper[4802]: I1125 17:03:46.003668 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e6a7e14-e3b5-4105-8efb-25f2fd871dd6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:03:46 crc kubenswrapper[4802]: I1125 17:03:46.003705 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7kms\" (UniqueName: \"kubernetes.io/projected/9e6a7e14-e3b5-4105-8efb-25f2fd871dd6-kube-api-access-b7kms\") on node \"crc\" DevicePath \"\"" Nov 25 17:03:46 crc kubenswrapper[4802]: I1125 17:03:46.490447 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-6222-account-create-update-qzmx5" event={"ID":"9e6a7e14-e3b5-4105-8efb-25f2fd871dd6","Type":"ContainerDied","Data":"9cf2221b6a59eea93f47c27d7b1e70df90e8a6a9475f073a92a2a526658a5049"} Nov 25 17:03:46 crc kubenswrapper[4802]: I1125 17:03:46.490497 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-6222-account-create-update-qzmx5" Nov 25 17:03:46 crc kubenswrapper[4802]: I1125 17:03:46.490507 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cf2221b6a59eea93f47c27d7b1e70df90e8a6a9475f073a92a2a526658a5049" Nov 25 17:03:46 crc kubenswrapper[4802]: I1125 17:03:46.694592 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-ln4wf"] Nov 25 17:03:46 crc kubenswrapper[4802]: E1125 17:03:46.695281 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2004be61-2b02-4bc4-a79e-5beba8cb4a60" containerName="mariadb-database-create" Nov 25 17:03:46 crc kubenswrapper[4802]: I1125 17:03:46.695323 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="2004be61-2b02-4bc4-a79e-5beba8cb4a60" containerName="mariadb-database-create" Nov 25 17:03:46 crc kubenswrapper[4802]: E1125 17:03:46.695337 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6a7e14-e3b5-4105-8efb-25f2fd871dd6" containerName="mariadb-account-create-update" Nov 25 17:03:46 crc kubenswrapper[4802]: I1125 17:03:46.695347 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6a7e14-e3b5-4105-8efb-25f2fd871dd6" containerName="mariadb-account-create-update" Nov 25 17:03:46 crc kubenswrapper[4802]: I1125 17:03:46.695533 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="2004be61-2b02-4bc4-a79e-5beba8cb4a60" containerName="mariadb-database-create" Nov 25 17:03:46 crc kubenswrapper[4802]: I1125 17:03:46.695571 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e6a7e14-e3b5-4105-8efb-25f2fd871dd6" containerName="mariadb-account-create-update" Nov 25 17:03:46 crc kubenswrapper[4802]: I1125 17:03:46.696391 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-ln4wf" Nov 25 17:03:46 crc kubenswrapper[4802]: I1125 17:03:46.699065 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-xdw54" Nov 25 17:03:46 crc kubenswrapper[4802]: I1125 17:03:46.702251 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-ln4wf"] Nov 25 17:03:46 crc kubenswrapper[4802]: I1125 17:03:46.814876 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87f7h\" (UniqueName: \"kubernetes.io/projected/f2307267-387c-4796-8971-4e79a1be93e9-kube-api-access-87f7h\") pod \"swift-operator-index-ln4wf\" (UID: \"f2307267-387c-4796-8971-4e79a1be93e9\") " pod="openstack-operators/swift-operator-index-ln4wf" Nov 25 17:03:46 crc kubenswrapper[4802]: I1125 17:03:46.915889 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87f7h\" (UniqueName: \"kubernetes.io/projected/f2307267-387c-4796-8971-4e79a1be93e9-kube-api-access-87f7h\") pod \"swift-operator-index-ln4wf\" (UID: \"f2307267-387c-4796-8971-4e79a1be93e9\") " pod="openstack-operators/swift-operator-index-ln4wf" Nov 25 17:03:46 crc kubenswrapper[4802]: I1125 17:03:46.936601 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87f7h\" (UniqueName: \"kubernetes.io/projected/f2307267-387c-4796-8971-4e79a1be93e9-kube-api-access-87f7h\") pod \"swift-operator-index-ln4wf\" (UID: \"f2307267-387c-4796-8971-4e79a1be93e9\") " pod="openstack-operators/swift-operator-index-ln4wf" Nov 25 17:03:47 crc kubenswrapper[4802]: I1125 17:03:47.014272 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-ln4wf" Nov 25 17:03:47 crc kubenswrapper[4802]: I1125 17:03:47.414329 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-ln4wf"] Nov 25 17:03:47 crc kubenswrapper[4802]: W1125 17:03:47.420356 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2307267_387c_4796_8971_4e79a1be93e9.slice/crio-e9bc00815301d268b1696b9a3c484056a93b110fc0c7b6c4eed5e31d240b27b9 WatchSource:0}: Error finding container e9bc00815301d268b1696b9a3c484056a93b110fc0c7b6c4eed5e31d240b27b9: Status 404 returned error can't find the container with id e9bc00815301d268b1696b9a3c484056a93b110fc0c7b6c4eed5e31d240b27b9 Nov 25 17:03:47 crc kubenswrapper[4802]: I1125 17:03:47.496909 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-ln4wf" event={"ID":"f2307267-387c-4796-8971-4e79a1be93e9","Type":"ContainerStarted","Data":"e9bc00815301d268b1696b9a3c484056a93b110fc0c7b6c4eed5e31d240b27b9"} Nov 25 17:03:49 crc kubenswrapper[4802]: I1125 17:03:49.488494 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/horizon-operator-index-xrnj5"] Nov 25 17:03:49 crc kubenswrapper[4802]: I1125 17:03:49.489361 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/horizon-operator-index-xrnj5" podUID="750a14b5-f70f-4945-9e4a-09e4e527ae62" containerName="registry-server" containerID="cri-o://a1cd5e9ab4ba82e1659afe7f8372e313322719dae0997cee11960b4bc99bb5b3" gracePeriod=2 Nov 25 17:03:50 crc kubenswrapper[4802]: I1125 17:03:50.297979 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-dt6d7"] Nov 25 17:03:50 crc kubenswrapper[4802]: I1125 17:03:50.298894 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-dt6d7" Nov 25 17:03:50 crc kubenswrapper[4802]: I1125 17:03:50.308178 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-dt6d7"] Nov 25 17:03:50 crc kubenswrapper[4802]: E1125 17:03:50.364951 4802 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod750a14b5_f70f_4945_9e4a_09e4e527ae62.slice/crio-conmon-a1cd5e9ab4ba82e1659afe7f8372e313322719dae0997cee11960b4bc99bb5b3.scope\": RecentStats: unable to find data in memory cache]" Nov 25 17:03:50 crc kubenswrapper[4802]: I1125 17:03:50.465707 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsw6x\" (UniqueName: \"kubernetes.io/projected/bce25e4a-fd42-4ffd-9c42-fad1e41af5e6-kube-api-access-bsw6x\") pod \"horizon-operator-index-dt6d7\" (UID: \"bce25e4a-fd42-4ffd-9c42-fad1e41af5e6\") " pod="openstack-operators/horizon-operator-index-dt6d7" Nov 25 17:03:50 crc kubenswrapper[4802]: I1125 17:03:50.532041 4802 generic.go:334] "Generic (PLEG): container finished" podID="750a14b5-f70f-4945-9e4a-09e4e527ae62" containerID="a1cd5e9ab4ba82e1659afe7f8372e313322719dae0997cee11960b4bc99bb5b3" exitCode=0 Nov 25 17:03:50 crc kubenswrapper[4802]: I1125 17:03:50.532082 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-xrnj5" event={"ID":"750a14b5-f70f-4945-9e4a-09e4e527ae62","Type":"ContainerDied","Data":"a1cd5e9ab4ba82e1659afe7f8372e313322719dae0997cee11960b4bc99bb5b3"} Nov 25 17:03:50 crc kubenswrapper[4802]: I1125 17:03:50.568194 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsw6x\" (UniqueName: \"kubernetes.io/projected/bce25e4a-fd42-4ffd-9c42-fad1e41af5e6-kube-api-access-bsw6x\") pod \"horizon-operator-index-dt6d7\" (UID: \"bce25e4a-fd42-4ffd-9c42-fad1e41af5e6\") " pod="openstack-operators/horizon-operator-index-dt6d7" Nov 25 17:03:50 crc kubenswrapper[4802]: I1125 17:03:50.579340 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-xrnj5" Nov 25 17:03:50 crc kubenswrapper[4802]: I1125 17:03:50.589686 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsw6x\" (UniqueName: \"kubernetes.io/projected/bce25e4a-fd42-4ffd-9c42-fad1e41af5e6-kube-api-access-bsw6x\") pod \"horizon-operator-index-dt6d7\" (UID: \"bce25e4a-fd42-4ffd-9c42-fad1e41af5e6\") " pod="openstack-operators/horizon-operator-index-dt6d7" Nov 25 17:03:50 crc kubenswrapper[4802]: I1125 17:03:50.668696 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tq9c\" (UniqueName: \"kubernetes.io/projected/750a14b5-f70f-4945-9e4a-09e4e527ae62-kube-api-access-5tq9c\") pod \"750a14b5-f70f-4945-9e4a-09e4e527ae62\" (UID: \"750a14b5-f70f-4945-9e4a-09e4e527ae62\") " Nov 25 17:03:50 crc kubenswrapper[4802]: I1125 17:03:50.672770 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/750a14b5-f70f-4945-9e4a-09e4e527ae62-kube-api-access-5tq9c" (OuterVolumeSpecName: "kube-api-access-5tq9c") pod "750a14b5-f70f-4945-9e4a-09e4e527ae62" (UID: "750a14b5-f70f-4945-9e4a-09e4e527ae62"). InnerVolumeSpecName "kube-api-access-5tq9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:03:50 crc kubenswrapper[4802]: I1125 17:03:50.672942 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-dt6d7" Nov 25 17:03:50 crc kubenswrapper[4802]: I1125 17:03:50.770726 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tq9c\" (UniqueName: \"kubernetes.io/projected/750a14b5-f70f-4945-9e4a-09e4e527ae62-kube-api-access-5tq9c\") on node \"crc\" DevicePath \"\"" Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.076935 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-dt6d7"] Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.278041 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.539312 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-dt6d7" event={"ID":"bce25e4a-fd42-4ffd-9c42-fad1e41af5e6","Type":"ContainerStarted","Data":"c1c53251d9acd2bb75c59179c492d7628ad1dc3d01c8f86f7dfa2b9e7c2c4379"} Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.540993 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-xrnj5" event={"ID":"750a14b5-f70f-4945-9e4a-09e4e527ae62","Type":"ContainerDied","Data":"c5f2969d40e26a177ba2f0fc0e4c1bf75c74004050a4a9508db343755bc5424c"} Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.541010 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-xrnj5" Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.541043 4802 scope.go:117] "RemoveContainer" containerID="a1cd5e9ab4ba82e1659afe7f8372e313322719dae0997cee11960b4bc99bb5b3" Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.542527 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-ln4wf" event={"ID":"f2307267-387c-4796-8971-4e79a1be93e9","Type":"ContainerStarted","Data":"6fd243ea5ee299e6325cb4a623777fa6e252cb3516595e21332d614198f5f338"} Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.558855 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-ln4wf" podStartSLOduration=2.597937877 podStartE2EDuration="5.558832809s" podCreationTimestamp="2025-11-25 17:03:46 +0000 UTC" firstStartedPulling="2025-11-25 17:03:47.422759555 +0000 UTC m=+1010.567106731" lastFinishedPulling="2025-11-25 17:03:50.383654467 +0000 UTC m=+1013.528001663" observedRunningTime="2025-11-25 17:03:51.556327492 +0000 UTC m=+1014.700674688" watchObservedRunningTime="2025-11-25 17:03:51.558832809 +0000 UTC m=+1014.703179995" Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.577201 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/horizon-operator-index-xrnj5"] Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.581480 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/horizon-operator-index-xrnj5"] Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.884463 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-sync-7ghgc"] Nov 25 17:03:51 crc kubenswrapper[4802]: E1125 17:03:51.884715 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="750a14b5-f70f-4945-9e4a-09e4e527ae62" containerName="registry-server" Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.884727 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="750a14b5-f70f-4945-9e4a-09e4e527ae62" containerName="registry-server" Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.884840 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="750a14b5-f70f-4945-9e4a-09e4e527ae62" containerName="registry-server" Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.885344 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-7ghgc" Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.888419 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.888701 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-rq8mq" Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.888988 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.889289 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.903481 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-7ghgc"] Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.987173 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75eae88a-5232-4493-9fa1-895961fb3d26-config-data\") pod \"keystone-db-sync-7ghgc\" (UID: \"75eae88a-5232-4493-9fa1-895961fb3d26\") " pod="glance-kuttl-tests/keystone-db-sync-7ghgc" Nov 25 17:03:51 crc kubenswrapper[4802]: I1125 17:03:51.987290 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdmkn\" (UniqueName: \"kubernetes.io/projected/75eae88a-5232-4493-9fa1-895961fb3d26-kube-api-access-qdmkn\") pod \"keystone-db-sync-7ghgc\" (UID: \"75eae88a-5232-4493-9fa1-895961fb3d26\") " pod="glance-kuttl-tests/keystone-db-sync-7ghgc" Nov 25 17:03:52 crc kubenswrapper[4802]: I1125 17:03:52.088885 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75eae88a-5232-4493-9fa1-895961fb3d26-config-data\") pod \"keystone-db-sync-7ghgc\" (UID: \"75eae88a-5232-4493-9fa1-895961fb3d26\") " pod="glance-kuttl-tests/keystone-db-sync-7ghgc" Nov 25 17:03:52 crc kubenswrapper[4802]: I1125 17:03:52.088999 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdmkn\" (UniqueName: \"kubernetes.io/projected/75eae88a-5232-4493-9fa1-895961fb3d26-kube-api-access-qdmkn\") pod \"keystone-db-sync-7ghgc\" (UID: \"75eae88a-5232-4493-9fa1-895961fb3d26\") " pod="glance-kuttl-tests/keystone-db-sync-7ghgc" Nov 25 17:03:52 crc kubenswrapper[4802]: I1125 17:03:52.095775 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75eae88a-5232-4493-9fa1-895961fb3d26-config-data\") pod \"keystone-db-sync-7ghgc\" (UID: \"75eae88a-5232-4493-9fa1-895961fb3d26\") " pod="glance-kuttl-tests/keystone-db-sync-7ghgc" Nov 25 17:03:52 crc kubenswrapper[4802]: I1125 17:03:52.105340 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdmkn\" (UniqueName: \"kubernetes.io/projected/75eae88a-5232-4493-9fa1-895961fb3d26-kube-api-access-qdmkn\") pod \"keystone-db-sync-7ghgc\" (UID: \"75eae88a-5232-4493-9fa1-895961fb3d26\") " pod="glance-kuttl-tests/keystone-db-sync-7ghgc" Nov 25 17:03:52 crc kubenswrapper[4802]: I1125 17:03:52.203555 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-7ghgc" Nov 25 17:03:52 crc kubenswrapper[4802]: I1125 17:03:52.550473 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-dt6d7" event={"ID":"bce25e4a-fd42-4ffd-9c42-fad1e41af5e6","Type":"ContainerStarted","Data":"c2ff5003ad8c9c7fa1a3828585d6c753247ca96bdacc6ebf3dac2a32ef176701"} Nov 25 17:03:52 crc kubenswrapper[4802]: I1125 17:03:52.566591 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-dt6d7" podStartSLOduration=1.929792157 podStartE2EDuration="2.566563856s" podCreationTimestamp="2025-11-25 17:03:50 +0000 UTC" firstStartedPulling="2025-11-25 17:03:51.088018696 +0000 UTC m=+1014.232365922" lastFinishedPulling="2025-11-25 17:03:51.724790435 +0000 UTC m=+1014.869137621" observedRunningTime="2025-11-25 17:03:52.565188658 +0000 UTC m=+1015.709535844" watchObservedRunningTime="2025-11-25 17:03:52.566563856 +0000 UTC m=+1015.710911042" Nov 25 17:03:52 crc kubenswrapper[4802]: I1125 17:03:52.612406 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-7ghgc"] Nov 25 17:03:52 crc kubenswrapper[4802]: I1125 17:03:52.686665 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-index-ln4wf"] Nov 25 17:03:53 crc kubenswrapper[4802]: I1125 17:03:53.296906 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-hp7k9"] Nov 25 17:03:53 crc kubenswrapper[4802]: I1125 17:03:53.298062 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-hp7k9" Nov 25 17:03:53 crc kubenswrapper[4802]: I1125 17:03:53.310536 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-hp7k9"] Nov 25 17:03:53 crc kubenswrapper[4802]: I1125 17:03:53.406487 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llzpk\" (UniqueName: \"kubernetes.io/projected/459e8530-a5fb-4b88-b42d-b68cf4d41835-kube-api-access-llzpk\") pod \"swift-operator-index-hp7k9\" (UID: \"459e8530-a5fb-4b88-b42d-b68cf4d41835\") " pod="openstack-operators/swift-operator-index-hp7k9" Nov 25 17:03:53 crc kubenswrapper[4802]: I1125 17:03:53.507980 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llzpk\" (UniqueName: \"kubernetes.io/projected/459e8530-a5fb-4b88-b42d-b68cf4d41835-kube-api-access-llzpk\") pod \"swift-operator-index-hp7k9\" (UID: \"459e8530-a5fb-4b88-b42d-b68cf4d41835\") " pod="openstack-operators/swift-operator-index-hp7k9" Nov 25 17:03:53 crc kubenswrapper[4802]: I1125 17:03:53.515780 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="750a14b5-f70f-4945-9e4a-09e4e527ae62" path="/var/lib/kubelet/pods/750a14b5-f70f-4945-9e4a-09e4e527ae62/volumes" Nov 25 17:03:53 crc kubenswrapper[4802]: I1125 17:03:53.535774 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llzpk\" (UniqueName: \"kubernetes.io/projected/459e8530-a5fb-4b88-b42d-b68cf4d41835-kube-api-access-llzpk\") pod \"swift-operator-index-hp7k9\" (UID: \"459e8530-a5fb-4b88-b42d-b68cf4d41835\") " pod="openstack-operators/swift-operator-index-hp7k9" Nov 25 17:03:53 crc kubenswrapper[4802]: I1125 17:03:53.562052 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-7ghgc" event={"ID":"75eae88a-5232-4493-9fa1-895961fb3d26","Type":"ContainerStarted","Data":"7ab637544434edaba19059634ab816086457d105601981a0f015e00f94b3a55c"} Nov 25 17:03:53 crc kubenswrapper[4802]: I1125 17:03:53.562239 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/swift-operator-index-ln4wf" podUID="f2307267-387c-4796-8971-4e79a1be93e9" containerName="registry-server" containerID="cri-o://6fd243ea5ee299e6325cb4a623777fa6e252cb3516595e21332d614198f5f338" gracePeriod=2 Nov 25 17:03:53 crc kubenswrapper[4802]: I1125 17:03:53.621716 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-hp7k9" Nov 25 17:03:54 crc kubenswrapper[4802]: I1125 17:03:54.039787 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-ln4wf" Nov 25 17:03:54 crc kubenswrapper[4802]: I1125 17:03:54.076519 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-hp7k9"] Nov 25 17:03:54 crc kubenswrapper[4802]: W1125 17:03:54.087535 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod459e8530_a5fb_4b88_b42d_b68cf4d41835.slice/crio-6c40cac1575f986f6c9ddaf6674547d31b8cd68a5dbae5f8d35564da6c9cfe94 WatchSource:0}: Error finding container 6c40cac1575f986f6c9ddaf6674547d31b8cd68a5dbae5f8d35564da6c9cfe94: Status 404 returned error can't find the container with id 6c40cac1575f986f6c9ddaf6674547d31b8cd68a5dbae5f8d35564da6c9cfe94 Nov 25 17:03:54 crc kubenswrapper[4802]: I1125 17:03:54.118682 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87f7h\" (UniqueName: \"kubernetes.io/projected/f2307267-387c-4796-8971-4e79a1be93e9-kube-api-access-87f7h\") pod \"f2307267-387c-4796-8971-4e79a1be93e9\" (UID: \"f2307267-387c-4796-8971-4e79a1be93e9\") " Nov 25 17:03:54 crc kubenswrapper[4802]: I1125 17:03:54.127187 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2307267-387c-4796-8971-4e79a1be93e9-kube-api-access-87f7h" (OuterVolumeSpecName: "kube-api-access-87f7h") pod "f2307267-387c-4796-8971-4e79a1be93e9" (UID: "f2307267-387c-4796-8971-4e79a1be93e9"). InnerVolumeSpecName "kube-api-access-87f7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:03:54 crc kubenswrapper[4802]: I1125 17:03:54.221095 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87f7h\" (UniqueName: \"kubernetes.io/projected/f2307267-387c-4796-8971-4e79a1be93e9-kube-api-access-87f7h\") on node \"crc\" DevicePath \"\"" Nov 25 17:03:54 crc kubenswrapper[4802]: I1125 17:03:54.572501 4802 generic.go:334] "Generic (PLEG): container finished" podID="f2307267-387c-4796-8971-4e79a1be93e9" containerID="6fd243ea5ee299e6325cb4a623777fa6e252cb3516595e21332d614198f5f338" exitCode=0 Nov 25 17:03:54 crc kubenswrapper[4802]: I1125 17:03:54.572827 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-ln4wf" Nov 25 17:03:54 crc kubenswrapper[4802]: I1125 17:03:54.573254 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-ln4wf" event={"ID":"f2307267-387c-4796-8971-4e79a1be93e9","Type":"ContainerDied","Data":"6fd243ea5ee299e6325cb4a623777fa6e252cb3516595e21332d614198f5f338"} Nov 25 17:03:54 crc kubenswrapper[4802]: I1125 17:03:54.573461 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-ln4wf" event={"ID":"f2307267-387c-4796-8971-4e79a1be93e9","Type":"ContainerDied","Data":"e9bc00815301d268b1696b9a3c484056a93b110fc0c7b6c4eed5e31d240b27b9"} Nov 25 17:03:54 crc kubenswrapper[4802]: I1125 17:03:54.573507 4802 scope.go:117] "RemoveContainer" containerID="6fd243ea5ee299e6325cb4a623777fa6e252cb3516595e21332d614198f5f338" Nov 25 17:03:54 crc kubenswrapper[4802]: I1125 17:03:54.574969 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-hp7k9" event={"ID":"459e8530-a5fb-4b88-b42d-b68cf4d41835","Type":"ContainerStarted","Data":"6c40cac1575f986f6c9ddaf6674547d31b8cd68a5dbae5f8d35564da6c9cfe94"} Nov 25 17:03:54 crc kubenswrapper[4802]: I1125 17:03:54.601587 4802 scope.go:117] "RemoveContainer" containerID="6fd243ea5ee299e6325cb4a623777fa6e252cb3516595e21332d614198f5f338" Nov 25 17:03:54 crc kubenswrapper[4802]: E1125 17:03:54.601991 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fd243ea5ee299e6325cb4a623777fa6e252cb3516595e21332d614198f5f338\": container with ID starting with 6fd243ea5ee299e6325cb4a623777fa6e252cb3516595e21332d614198f5f338 not found: ID does not exist" containerID="6fd243ea5ee299e6325cb4a623777fa6e252cb3516595e21332d614198f5f338" Nov 25 17:03:54 crc kubenswrapper[4802]: I1125 17:03:54.603915 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fd243ea5ee299e6325cb4a623777fa6e252cb3516595e21332d614198f5f338"} err="failed to get container status \"6fd243ea5ee299e6325cb4a623777fa6e252cb3516595e21332d614198f5f338\": rpc error: code = NotFound desc = could not find container \"6fd243ea5ee299e6325cb4a623777fa6e252cb3516595e21332d614198f5f338\": container with ID starting with 6fd243ea5ee299e6325cb4a623777fa6e252cb3516595e21332d614198f5f338 not found: ID does not exist" Nov 25 17:03:54 crc kubenswrapper[4802]: I1125 17:03:54.611945 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-index-ln4wf"] Nov 25 17:03:54 crc kubenswrapper[4802]: I1125 17:03:54.615930 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/swift-operator-index-ln4wf"] Nov 25 17:03:55 crc kubenswrapper[4802]: I1125 17:03:55.512326 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2307267-387c-4796-8971-4e79a1be93e9" path="/var/lib/kubelet/pods/f2307267-387c-4796-8971-4e79a1be93e9/volumes" Nov 25 17:04:00 crc kubenswrapper[4802]: I1125 17:04:00.673093 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-index-dt6d7" Nov 25 17:04:00 crc kubenswrapper[4802]: I1125 17:04:00.673779 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/horizon-operator-index-dt6d7" Nov 25 17:04:00 crc kubenswrapper[4802]: I1125 17:04:00.707083 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/horizon-operator-index-dt6d7" Nov 25 17:04:01 crc kubenswrapper[4802]: I1125 17:04:01.662438 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-index-dt6d7" Nov 25 17:04:03 crc kubenswrapper[4802]: I1125 17:04:03.638462 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-hp7k9" event={"ID":"459e8530-a5fb-4b88-b42d-b68cf4d41835","Type":"ContainerStarted","Data":"427b817ca13a6a0458f60935160dec959e784a213a326041a1b1c8cca80a218a"} Nov 25 17:04:03 crc kubenswrapper[4802]: I1125 17:04:03.639978 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-7ghgc" event={"ID":"75eae88a-5232-4493-9fa1-895961fb3d26","Type":"ContainerStarted","Data":"8cec2010a2a8ba7ea562b05c08b0877d5544ac04385e73e610ffa50e4bc88dbe"} Nov 25 17:04:03 crc kubenswrapper[4802]: I1125 17:04:03.655112 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-hp7k9" podStartSLOduration=1.3948554180000001 podStartE2EDuration="10.655091821s" podCreationTimestamp="2025-11-25 17:03:53 +0000 UTC" firstStartedPulling="2025-11-25 17:03:54.093088464 +0000 UTC m=+1017.237435650" lastFinishedPulling="2025-11-25 17:04:03.353324867 +0000 UTC m=+1026.497672053" observedRunningTime="2025-11-25 17:04:03.65396175 +0000 UTC m=+1026.798308946" watchObservedRunningTime="2025-11-25 17:04:03.655091821 +0000 UTC m=+1026.799439007" Nov 25 17:04:03 crc kubenswrapper[4802]: I1125 17:04:03.675979 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-db-sync-7ghgc" podStartSLOduration=1.9495111189999998 podStartE2EDuration="12.675961407s" podCreationTimestamp="2025-11-25 17:03:51 +0000 UTC" firstStartedPulling="2025-11-25 17:03:52.623536759 +0000 UTC m=+1015.767883945" lastFinishedPulling="2025-11-25 17:04:03.349987047 +0000 UTC m=+1026.494334233" observedRunningTime="2025-11-25 17:04:03.673691435 +0000 UTC m=+1026.818038611" watchObservedRunningTime="2025-11-25 17:04:03.675961407 +0000 UTC m=+1026.820308593" Nov 25 17:04:09 crc kubenswrapper[4802]: I1125 17:04:09.683266 4802 generic.go:334] "Generic (PLEG): container finished" podID="75eae88a-5232-4493-9fa1-895961fb3d26" containerID="8cec2010a2a8ba7ea562b05c08b0877d5544ac04385e73e610ffa50e4bc88dbe" exitCode=0 Nov 25 17:04:09 crc kubenswrapper[4802]: I1125 17:04:09.683397 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-7ghgc" event={"ID":"75eae88a-5232-4493-9fa1-895961fb3d26","Type":"ContainerDied","Data":"8cec2010a2a8ba7ea562b05c08b0877d5544ac04385e73e610ffa50e4bc88dbe"} Nov 25 17:04:10 crc kubenswrapper[4802]: I1125 17:04:10.977085 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-7ghgc" Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.088168 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75eae88a-5232-4493-9fa1-895961fb3d26-config-data\") pod \"75eae88a-5232-4493-9fa1-895961fb3d26\" (UID: \"75eae88a-5232-4493-9fa1-895961fb3d26\") " Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.088237 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdmkn\" (UniqueName: \"kubernetes.io/projected/75eae88a-5232-4493-9fa1-895961fb3d26-kube-api-access-qdmkn\") pod \"75eae88a-5232-4493-9fa1-895961fb3d26\" (UID: \"75eae88a-5232-4493-9fa1-895961fb3d26\") " Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.094888 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75eae88a-5232-4493-9fa1-895961fb3d26-kube-api-access-qdmkn" (OuterVolumeSpecName: "kube-api-access-qdmkn") pod "75eae88a-5232-4493-9fa1-895961fb3d26" (UID: "75eae88a-5232-4493-9fa1-895961fb3d26"). InnerVolumeSpecName "kube-api-access-qdmkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.127563 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75eae88a-5232-4493-9fa1-895961fb3d26-config-data" (OuterVolumeSpecName: "config-data") pod "75eae88a-5232-4493-9fa1-895961fb3d26" (UID: "75eae88a-5232-4493-9fa1-895961fb3d26"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.190664 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75eae88a-5232-4493-9fa1-895961fb3d26-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.190750 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdmkn\" (UniqueName: \"kubernetes.io/projected/75eae88a-5232-4493-9fa1-895961fb3d26-kube-api-access-qdmkn\") on node \"crc\" DevicePath \"\"" Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.698601 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-7ghgc" event={"ID":"75eae88a-5232-4493-9fa1-895961fb3d26","Type":"ContainerDied","Data":"7ab637544434edaba19059634ab816086457d105601981a0f015e00f94b3a55c"} Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.698638 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ab637544434edaba19059634ab816086457d105601981a0f015e00f94b3a55c" Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.698687 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-7ghgc" Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.898950 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-7czbg"] Nov 25 17:04:11 crc kubenswrapper[4802]: E1125 17:04:11.899256 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75eae88a-5232-4493-9fa1-895961fb3d26" containerName="keystone-db-sync" Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.899273 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="75eae88a-5232-4493-9fa1-895961fb3d26" containerName="keystone-db-sync" Nov 25 17:04:11 crc kubenswrapper[4802]: E1125 17:04:11.899288 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2307267-387c-4796-8971-4e79a1be93e9" containerName="registry-server" Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.899295 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2307267-387c-4796-8971-4e79a1be93e9" containerName="registry-server" Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.899404 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="75eae88a-5232-4493-9fa1-895961fb3d26" containerName="keystone-db-sync" Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.899417 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2307267-387c-4796-8971-4e79a1be93e9" containerName="registry-server" Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.899831 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.901996 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.902318 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"osp-secret" Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.902471 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.902485 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.903846 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-rq8mq" Nov 25 17:04:11 crc kubenswrapper[4802]: I1125 17:04:11.928302 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-7czbg"] Nov 25 17:04:12 crc kubenswrapper[4802]: I1125 17:04:12.001966 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-fernet-keys\") pod \"keystone-bootstrap-7czbg\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:12 crc kubenswrapper[4802]: I1125 17:04:12.002096 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-scripts\") pod \"keystone-bootstrap-7czbg\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:12 crc kubenswrapper[4802]: I1125 17:04:12.002143 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-credential-keys\") pod \"keystone-bootstrap-7czbg\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:12 crc kubenswrapper[4802]: I1125 17:04:12.002168 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5bkg\" (UniqueName: \"kubernetes.io/projected/d9278966-eb40-4107-b492-2012500b1701-kube-api-access-j5bkg\") pod \"keystone-bootstrap-7czbg\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:12 crc kubenswrapper[4802]: I1125 17:04:12.002193 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-config-data\") pod \"keystone-bootstrap-7czbg\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:12 crc kubenswrapper[4802]: I1125 17:04:12.104035 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-fernet-keys\") pod \"keystone-bootstrap-7czbg\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:12 crc kubenswrapper[4802]: I1125 17:04:12.104176 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-scripts\") pod \"keystone-bootstrap-7czbg\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:12 crc kubenswrapper[4802]: I1125 17:04:12.104201 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-credential-keys\") pod \"keystone-bootstrap-7czbg\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:12 crc kubenswrapper[4802]: I1125 17:04:12.104223 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5bkg\" (UniqueName: \"kubernetes.io/projected/d9278966-eb40-4107-b492-2012500b1701-kube-api-access-j5bkg\") pod \"keystone-bootstrap-7czbg\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:12 crc kubenswrapper[4802]: I1125 17:04:12.104249 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-config-data\") pod \"keystone-bootstrap-7czbg\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:12 crc kubenswrapper[4802]: I1125 17:04:12.109023 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-fernet-keys\") pod \"keystone-bootstrap-7czbg\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:12 crc kubenswrapper[4802]: I1125 17:04:12.109249 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-config-data\") pod \"keystone-bootstrap-7czbg\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:12 crc kubenswrapper[4802]: I1125 17:04:12.109354 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-scripts\") pod \"keystone-bootstrap-7czbg\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:12 crc kubenswrapper[4802]: I1125 17:04:12.111373 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-credential-keys\") pod \"keystone-bootstrap-7czbg\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:12 crc kubenswrapper[4802]: I1125 17:04:12.122834 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5bkg\" (UniqueName: \"kubernetes.io/projected/d9278966-eb40-4107-b492-2012500b1701-kube-api-access-j5bkg\") pod \"keystone-bootstrap-7czbg\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:12 crc kubenswrapper[4802]: I1125 17:04:12.234858 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:12 crc kubenswrapper[4802]: I1125 17:04:12.639209 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-7czbg"] Nov 25 17:04:12 crc kubenswrapper[4802]: I1125 17:04:12.709592 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-7czbg" event={"ID":"d9278966-eb40-4107-b492-2012500b1701","Type":"ContainerStarted","Data":"1a08dbb27616c3cff1b5021283fa37c17defaa26e255e94ac59d7f6ef406ecff"} Nov 25 17:04:13 crc kubenswrapper[4802]: I1125 17:04:13.622810 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-hp7k9" Nov 25 17:04:13 crc kubenswrapper[4802]: I1125 17:04:13.622873 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-hp7k9" Nov 25 17:04:13 crc kubenswrapper[4802]: I1125 17:04:13.659068 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-hp7k9" Nov 25 17:04:13 crc kubenswrapper[4802]: I1125 17:04:13.719990 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-7czbg" event={"ID":"d9278966-eb40-4107-b492-2012500b1701","Type":"ContainerStarted","Data":"162bdc1af7a8245196ce544a3deff91163dd4f10dd7669e0b6a1ab8a42921fad"} Nov 25 17:04:13 crc kubenswrapper[4802]: I1125 17:04:13.746424 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-bootstrap-7czbg" podStartSLOduration=2.746398584 podStartE2EDuration="2.746398584s" podCreationTimestamp="2025-11-25 17:04:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:04:13.738938022 +0000 UTC m=+1036.883285208" watchObservedRunningTime="2025-11-25 17:04:13.746398584 +0000 UTC m=+1036.890745770" Nov 25 17:04:13 crc kubenswrapper[4802]: I1125 17:04:13.749555 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-hp7k9" Nov 25 17:04:16 crc kubenswrapper[4802]: I1125 17:04:16.740191 4802 generic.go:334] "Generic (PLEG): container finished" podID="d9278966-eb40-4107-b492-2012500b1701" containerID="162bdc1af7a8245196ce544a3deff91163dd4f10dd7669e0b6a1ab8a42921fad" exitCode=0 Nov 25 17:04:16 crc kubenswrapper[4802]: I1125 17:04:16.740281 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-7czbg" event={"ID":"d9278966-eb40-4107-b492-2012500b1701","Type":"ContainerDied","Data":"162bdc1af7a8245196ce544a3deff91163dd4f10dd7669e0b6a1ab8a42921fad"} Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.002393 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.094740 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5bkg\" (UniqueName: \"kubernetes.io/projected/d9278966-eb40-4107-b492-2012500b1701-kube-api-access-j5bkg\") pod \"d9278966-eb40-4107-b492-2012500b1701\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.094793 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-scripts\") pod \"d9278966-eb40-4107-b492-2012500b1701\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.094809 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-credential-keys\") pod \"d9278966-eb40-4107-b492-2012500b1701\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.094875 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-config-data\") pod \"d9278966-eb40-4107-b492-2012500b1701\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.094903 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-fernet-keys\") pod \"d9278966-eb40-4107-b492-2012500b1701\" (UID: \"d9278966-eb40-4107-b492-2012500b1701\") " Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.101239 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d9278966-eb40-4107-b492-2012500b1701" (UID: "d9278966-eb40-4107-b492-2012500b1701"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.101492 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9278966-eb40-4107-b492-2012500b1701-kube-api-access-j5bkg" (OuterVolumeSpecName: "kube-api-access-j5bkg") pod "d9278966-eb40-4107-b492-2012500b1701" (UID: "d9278966-eb40-4107-b492-2012500b1701"). InnerVolumeSpecName "kube-api-access-j5bkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.102899 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-scripts" (OuterVolumeSpecName: "scripts") pod "d9278966-eb40-4107-b492-2012500b1701" (UID: "d9278966-eb40-4107-b492-2012500b1701"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.104091 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d9278966-eb40-4107-b492-2012500b1701" (UID: "d9278966-eb40-4107-b492-2012500b1701"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.115939 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-config-data" (OuterVolumeSpecName: "config-data") pod "d9278966-eb40-4107-b492-2012500b1701" (UID: "d9278966-eb40-4107-b492-2012500b1701"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.196386 4802 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.196427 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5bkg\" (UniqueName: \"kubernetes.io/projected/d9278966-eb40-4107-b492-2012500b1701-kube-api-access-j5bkg\") on node \"crc\" DevicePath \"\"" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.196442 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.196452 4802 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.196460 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9278966-eb40-4107-b492-2012500b1701-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.754584 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-7czbg" event={"ID":"d9278966-eb40-4107-b492-2012500b1701","Type":"ContainerDied","Data":"1a08dbb27616c3cff1b5021283fa37c17defaa26e255e94ac59d7f6ef406ecff"} Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.754616 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-7czbg" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.754630 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a08dbb27616c3cff1b5021283fa37c17defaa26e255e94ac59d7f6ef406ecff" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.834723 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-64bc67949c-jgvwk"] Nov 25 17:04:18 crc kubenswrapper[4802]: E1125 17:04:18.835228 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9278966-eb40-4107-b492-2012500b1701" containerName="keystone-bootstrap" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.835249 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9278966-eb40-4107-b492-2012500b1701" containerName="keystone-bootstrap" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.835436 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9278966-eb40-4107-b492-2012500b1701" containerName="keystone-bootstrap" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.836211 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.838989 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.839058 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.839073 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.839803 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-rq8mq" Nov 25 17:04:18 crc kubenswrapper[4802]: I1125 17:04:18.846853 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-64bc67949c-jgvwk"] Nov 25 17:04:19 crc kubenswrapper[4802]: I1125 17:04:19.010380 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b799cfff-4f77-4b6c-928e-0bcbed6fc523-fernet-keys\") pod \"keystone-64bc67949c-jgvwk\" (UID: \"b799cfff-4f77-4b6c-928e-0bcbed6fc523\") " pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:19 crc kubenswrapper[4802]: I1125 17:04:19.010476 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ml7c7\" (UniqueName: \"kubernetes.io/projected/b799cfff-4f77-4b6c-928e-0bcbed6fc523-kube-api-access-ml7c7\") pod \"keystone-64bc67949c-jgvwk\" (UID: \"b799cfff-4f77-4b6c-928e-0bcbed6fc523\") " pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:19 crc kubenswrapper[4802]: I1125 17:04:19.010597 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b799cfff-4f77-4b6c-928e-0bcbed6fc523-scripts\") pod \"keystone-64bc67949c-jgvwk\" (UID: \"b799cfff-4f77-4b6c-928e-0bcbed6fc523\") " pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:19 crc kubenswrapper[4802]: I1125 17:04:19.010648 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b799cfff-4f77-4b6c-928e-0bcbed6fc523-credential-keys\") pod \"keystone-64bc67949c-jgvwk\" (UID: \"b799cfff-4f77-4b6c-928e-0bcbed6fc523\") " pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:19 crc kubenswrapper[4802]: I1125 17:04:19.010793 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b799cfff-4f77-4b6c-928e-0bcbed6fc523-config-data\") pod \"keystone-64bc67949c-jgvwk\" (UID: \"b799cfff-4f77-4b6c-928e-0bcbed6fc523\") " pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:19 crc kubenswrapper[4802]: I1125 17:04:19.112100 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b799cfff-4f77-4b6c-928e-0bcbed6fc523-fernet-keys\") pod \"keystone-64bc67949c-jgvwk\" (UID: \"b799cfff-4f77-4b6c-928e-0bcbed6fc523\") " pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:19 crc kubenswrapper[4802]: I1125 17:04:19.112180 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml7c7\" (UniqueName: \"kubernetes.io/projected/b799cfff-4f77-4b6c-928e-0bcbed6fc523-kube-api-access-ml7c7\") pod \"keystone-64bc67949c-jgvwk\" (UID: \"b799cfff-4f77-4b6c-928e-0bcbed6fc523\") " pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:19 crc kubenswrapper[4802]: I1125 17:04:19.112244 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b799cfff-4f77-4b6c-928e-0bcbed6fc523-scripts\") pod \"keystone-64bc67949c-jgvwk\" (UID: \"b799cfff-4f77-4b6c-928e-0bcbed6fc523\") " pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:19 crc kubenswrapper[4802]: I1125 17:04:19.112275 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b799cfff-4f77-4b6c-928e-0bcbed6fc523-credential-keys\") pod \"keystone-64bc67949c-jgvwk\" (UID: \"b799cfff-4f77-4b6c-928e-0bcbed6fc523\") " pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:19 crc kubenswrapper[4802]: I1125 17:04:19.112327 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b799cfff-4f77-4b6c-928e-0bcbed6fc523-config-data\") pod \"keystone-64bc67949c-jgvwk\" (UID: \"b799cfff-4f77-4b6c-928e-0bcbed6fc523\") " pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:19 crc kubenswrapper[4802]: I1125 17:04:19.117774 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b799cfff-4f77-4b6c-928e-0bcbed6fc523-scripts\") pod \"keystone-64bc67949c-jgvwk\" (UID: \"b799cfff-4f77-4b6c-928e-0bcbed6fc523\") " pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:19 crc kubenswrapper[4802]: I1125 17:04:19.117886 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b799cfff-4f77-4b6c-928e-0bcbed6fc523-credential-keys\") pod \"keystone-64bc67949c-jgvwk\" (UID: \"b799cfff-4f77-4b6c-928e-0bcbed6fc523\") " pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:19 crc kubenswrapper[4802]: I1125 17:04:19.118015 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b799cfff-4f77-4b6c-928e-0bcbed6fc523-fernet-keys\") pod \"keystone-64bc67949c-jgvwk\" (UID: \"b799cfff-4f77-4b6c-928e-0bcbed6fc523\") " pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:19 crc kubenswrapper[4802]: I1125 17:04:19.118778 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b799cfff-4f77-4b6c-928e-0bcbed6fc523-config-data\") pod \"keystone-64bc67949c-jgvwk\" (UID: \"b799cfff-4f77-4b6c-928e-0bcbed6fc523\") " pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:19 crc kubenswrapper[4802]: I1125 17:04:19.131345 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml7c7\" (UniqueName: \"kubernetes.io/projected/b799cfff-4f77-4b6c-928e-0bcbed6fc523-kube-api-access-ml7c7\") pod \"keystone-64bc67949c-jgvwk\" (UID: \"b799cfff-4f77-4b6c-928e-0bcbed6fc523\") " pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:19 crc kubenswrapper[4802]: I1125 17:04:19.153607 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:19 crc kubenswrapper[4802]: I1125 17:04:19.612434 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-64bc67949c-jgvwk"] Nov 25 17:04:19 crc kubenswrapper[4802]: W1125 17:04:19.618812 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb799cfff_4f77_4b6c_928e_0bcbed6fc523.slice/crio-6c42556d15801bc4cad1269f7190c937fb45bf18e77f1a8953589c3e54b186bd WatchSource:0}: Error finding container 6c42556d15801bc4cad1269f7190c937fb45bf18e77f1a8953589c3e54b186bd: Status 404 returned error can't find the container with id 6c42556d15801bc4cad1269f7190c937fb45bf18e77f1a8953589c3e54b186bd Nov 25 17:04:19 crc kubenswrapper[4802]: I1125 17:04:19.763270 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" event={"ID":"b799cfff-4f77-4b6c-928e-0bcbed6fc523","Type":"ContainerStarted","Data":"6c42556d15801bc4cad1269f7190c937fb45bf18e77f1a8953589c3e54b186bd"} Nov 25 17:04:20 crc kubenswrapper[4802]: I1125 17:04:20.787148 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" event={"ID":"b799cfff-4f77-4b6c-928e-0bcbed6fc523","Type":"ContainerStarted","Data":"86330fa4753a3c5c5bc5e70b209d799da9e840eeb6dc5f3ef4e667b4df29f2e6"} Nov 25 17:04:20 crc kubenswrapper[4802]: I1125 17:04:20.788030 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:20 crc kubenswrapper[4802]: I1125 17:04:20.832579 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" podStartSLOduration=2.832545406 podStartE2EDuration="2.832545406s" podCreationTimestamp="2025-11-25 17:04:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:04:20.810774717 +0000 UTC m=+1043.955121913" watchObservedRunningTime="2025-11-25 17:04:20.832545406 +0000 UTC m=+1043.976892602" Nov 25 17:04:26 crc kubenswrapper[4802]: I1125 17:04:26.375508 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr"] Nov 25 17:04:26 crc kubenswrapper[4802]: I1125 17:04:26.377149 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" Nov 25 17:04:26 crc kubenswrapper[4802]: I1125 17:04:26.379705 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-t7662" Nov 25 17:04:26 crc kubenswrapper[4802]: I1125 17:04:26.392520 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr"] Nov 25 17:04:26 crc kubenswrapper[4802]: I1125 17:04:26.521581 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d0befe17-9e51-457b-86c7-8831cb064339-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr\" (UID: \"d0befe17-9e51-457b-86c7-8831cb064339\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" Nov 25 17:04:26 crc kubenswrapper[4802]: I1125 17:04:26.521642 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d0befe17-9e51-457b-86c7-8831cb064339-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr\" (UID: \"d0befe17-9e51-457b-86c7-8831cb064339\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" Nov 25 17:04:26 crc kubenswrapper[4802]: I1125 17:04:26.521673 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njfk4\" (UniqueName: \"kubernetes.io/projected/d0befe17-9e51-457b-86c7-8831cb064339-kube-api-access-njfk4\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr\" (UID: \"d0befe17-9e51-457b-86c7-8831cb064339\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" Nov 25 17:04:26 crc kubenswrapper[4802]: I1125 17:04:26.622524 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d0befe17-9e51-457b-86c7-8831cb064339-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr\" (UID: \"d0befe17-9e51-457b-86c7-8831cb064339\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" Nov 25 17:04:26 crc kubenswrapper[4802]: I1125 17:04:26.622892 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d0befe17-9e51-457b-86c7-8831cb064339-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr\" (UID: \"d0befe17-9e51-457b-86c7-8831cb064339\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" Nov 25 17:04:26 crc kubenswrapper[4802]: I1125 17:04:26.622996 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njfk4\" (UniqueName: \"kubernetes.io/projected/d0befe17-9e51-457b-86c7-8831cb064339-kube-api-access-njfk4\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr\" (UID: \"d0befe17-9e51-457b-86c7-8831cb064339\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" Nov 25 17:04:26 crc kubenswrapper[4802]: I1125 17:04:26.623843 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d0befe17-9e51-457b-86c7-8831cb064339-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr\" (UID: \"d0befe17-9e51-457b-86c7-8831cb064339\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" Nov 25 17:04:26 crc kubenswrapper[4802]: I1125 17:04:26.623921 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d0befe17-9e51-457b-86c7-8831cb064339-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr\" (UID: \"d0befe17-9e51-457b-86c7-8831cb064339\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" Nov 25 17:04:26 crc kubenswrapper[4802]: I1125 17:04:26.643360 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njfk4\" (UniqueName: \"kubernetes.io/projected/d0befe17-9e51-457b-86c7-8831cb064339-kube-api-access-njfk4\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr\" (UID: \"d0befe17-9e51-457b-86c7-8831cb064339\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" Nov 25 17:04:26 crc kubenswrapper[4802]: I1125 17:04:26.694321 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" Nov 25 17:04:27 crc kubenswrapper[4802]: I1125 17:04:27.121261 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr"] Nov 25 17:04:27 crc kubenswrapper[4802]: I1125 17:04:27.330844 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7"] Nov 25 17:04:27 crc kubenswrapper[4802]: I1125 17:04:27.332298 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" Nov 25 17:04:27 crc kubenswrapper[4802]: I1125 17:04:27.343927 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7"] Nov 25 17:04:27 crc kubenswrapper[4802]: I1125 17:04:27.433025 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n82z\" (UniqueName: \"kubernetes.io/projected/5b0bee27-5168-4984-9943-c93299d3ac58-kube-api-access-6n82z\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7\" (UID: \"5b0bee27-5168-4984-9943-c93299d3ac58\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" Nov 25 17:04:27 crc kubenswrapper[4802]: I1125 17:04:27.433116 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5b0bee27-5168-4984-9943-c93299d3ac58-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7\" (UID: \"5b0bee27-5168-4984-9943-c93299d3ac58\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" Nov 25 17:04:27 crc kubenswrapper[4802]: I1125 17:04:27.433186 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5b0bee27-5168-4984-9943-c93299d3ac58-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7\" (UID: \"5b0bee27-5168-4984-9943-c93299d3ac58\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" Nov 25 17:04:27 crc kubenswrapper[4802]: I1125 17:04:27.534088 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5b0bee27-5168-4984-9943-c93299d3ac58-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7\" (UID: \"5b0bee27-5168-4984-9943-c93299d3ac58\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" Nov 25 17:04:27 crc kubenswrapper[4802]: I1125 17:04:27.534175 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n82z\" (UniqueName: \"kubernetes.io/projected/5b0bee27-5168-4984-9943-c93299d3ac58-kube-api-access-6n82z\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7\" (UID: \"5b0bee27-5168-4984-9943-c93299d3ac58\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" Nov 25 17:04:27 crc kubenswrapper[4802]: I1125 17:04:27.534229 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5b0bee27-5168-4984-9943-c93299d3ac58-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7\" (UID: \"5b0bee27-5168-4984-9943-c93299d3ac58\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" Nov 25 17:04:27 crc kubenswrapper[4802]: I1125 17:04:27.534549 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5b0bee27-5168-4984-9943-c93299d3ac58-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7\" (UID: \"5b0bee27-5168-4984-9943-c93299d3ac58\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" Nov 25 17:04:27 crc kubenswrapper[4802]: I1125 17:04:27.534625 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5b0bee27-5168-4984-9943-c93299d3ac58-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7\" (UID: \"5b0bee27-5168-4984-9943-c93299d3ac58\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" Nov 25 17:04:27 crc kubenswrapper[4802]: I1125 17:04:27.552913 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n82z\" (UniqueName: \"kubernetes.io/projected/5b0bee27-5168-4984-9943-c93299d3ac58-kube-api-access-6n82z\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7\" (UID: \"5b0bee27-5168-4984-9943-c93299d3ac58\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" Nov 25 17:04:27 crc kubenswrapper[4802]: I1125 17:04:27.649526 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" Nov 25 17:04:27 crc kubenswrapper[4802]: I1125 17:04:27.834862 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" event={"ID":"d0befe17-9e51-457b-86c7-8831cb064339","Type":"ContainerStarted","Data":"89bfc345d4346bde83cd3bef262687e962c6780ef0012c9cd62b7fb1c8c8f32f"} Nov 25 17:04:28 crc kubenswrapper[4802]: I1125 17:04:28.048686 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7"] Nov 25 17:04:28 crc kubenswrapper[4802]: I1125 17:04:28.842928 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" event={"ID":"5b0bee27-5168-4984-9943-c93299d3ac58","Type":"ContainerStarted","Data":"124c9bfe0f4f7bf1e0b076658079cb6ddb7c7599e5c9e4920a8b2f6996d973b3"} Nov 25 17:04:30 crc kubenswrapper[4802]: I1125 17:04:30.857208 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" event={"ID":"5b0bee27-5168-4984-9943-c93299d3ac58","Type":"ContainerStarted","Data":"4fca48738ee76d3dab58d05b2dd1fe1715f2116b3192c491c51de315500b376a"} Nov 25 17:04:30 crc kubenswrapper[4802]: I1125 17:04:30.858467 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" event={"ID":"d0befe17-9e51-457b-86c7-8831cb064339","Type":"ContainerStarted","Data":"e473e1215d574a8cff070599b8a89ff5d073a5b8b0f1f0df1714b4614c340c43"} Nov 25 17:04:31 crc kubenswrapper[4802]: E1125 17:04:31.100163 4802 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0befe17_9e51_457b_86c7_8831cb064339.slice/crio-conmon-e473e1215d574a8cff070599b8a89ff5d073a5b8b0f1f0df1714b4614c340c43.scope\": RecentStats: unable to find data in memory cache]" Nov 25 17:04:31 crc kubenswrapper[4802]: I1125 17:04:31.865790 4802 generic.go:334] "Generic (PLEG): container finished" podID="5b0bee27-5168-4984-9943-c93299d3ac58" containerID="4fca48738ee76d3dab58d05b2dd1fe1715f2116b3192c491c51de315500b376a" exitCode=0 Nov 25 17:04:31 crc kubenswrapper[4802]: I1125 17:04:31.866110 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" event={"ID":"5b0bee27-5168-4984-9943-c93299d3ac58","Type":"ContainerDied","Data":"4fca48738ee76d3dab58d05b2dd1fe1715f2116b3192c491c51de315500b376a"} Nov 25 17:04:31 crc kubenswrapper[4802]: I1125 17:04:31.868065 4802 generic.go:334] "Generic (PLEG): container finished" podID="d0befe17-9e51-457b-86c7-8831cb064339" containerID="e473e1215d574a8cff070599b8a89ff5d073a5b8b0f1f0df1714b4614c340c43" exitCode=0 Nov 25 17:04:31 crc kubenswrapper[4802]: I1125 17:04:31.868095 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" event={"ID":"d0befe17-9e51-457b-86c7-8831cb064339","Type":"ContainerDied","Data":"e473e1215d574a8cff070599b8a89ff5d073a5b8b0f1f0df1714b4614c340c43"} Nov 25 17:04:31 crc kubenswrapper[4802]: I1125 17:04:31.868938 4802 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 17:04:43 crc kubenswrapper[4802]: I1125 17:04:43.940567 4802 generic.go:334] "Generic (PLEG): container finished" podID="d0befe17-9e51-457b-86c7-8831cb064339" containerID="38b3801b61a1c6f7b938fd155c5fb2ab697c11808ad9ddea77a96c431c9709dc" exitCode=0 Nov 25 17:04:43 crc kubenswrapper[4802]: I1125 17:04:43.941088 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" event={"ID":"d0befe17-9e51-457b-86c7-8831cb064339","Type":"ContainerDied","Data":"38b3801b61a1c6f7b938fd155c5fb2ab697c11808ad9ddea77a96c431c9709dc"} Nov 25 17:04:43 crc kubenswrapper[4802]: I1125 17:04:43.942657 4802 generic.go:334] "Generic (PLEG): container finished" podID="5b0bee27-5168-4984-9943-c93299d3ac58" containerID="381a63597ef9d96c659586be097430eadf29ee77ec4956e986533d007f6674fc" exitCode=0 Nov 25 17:04:43 crc kubenswrapper[4802]: I1125 17:04:43.942686 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" event={"ID":"5b0bee27-5168-4984-9943-c93299d3ac58","Type":"ContainerDied","Data":"381a63597ef9d96c659586be097430eadf29ee77ec4956e986533d007f6674fc"} Nov 25 17:04:44 crc kubenswrapper[4802]: I1125 17:04:44.951302 4802 generic.go:334] "Generic (PLEG): container finished" podID="5b0bee27-5168-4984-9943-c93299d3ac58" containerID="0e3b7f8862ada272b7860d33aeea72b199c46586804669f264e8c91489a7778b" exitCode=0 Nov 25 17:04:44 crc kubenswrapper[4802]: I1125 17:04:44.951437 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" event={"ID":"5b0bee27-5168-4984-9943-c93299d3ac58","Type":"ContainerDied","Data":"0e3b7f8862ada272b7860d33aeea72b199c46586804669f264e8c91489a7778b"} Nov 25 17:04:44 crc kubenswrapper[4802]: I1125 17:04:44.953084 4802 generic.go:334] "Generic (PLEG): container finished" podID="d0befe17-9e51-457b-86c7-8831cb064339" containerID="b354cc7a25d89392cf3b1625417bc775b130c883db2276d86b452575c9483b7b" exitCode=0 Nov 25 17:04:44 crc kubenswrapper[4802]: I1125 17:04:44.953114 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" event={"ID":"d0befe17-9e51-457b-86c7-8831cb064339","Type":"ContainerDied","Data":"b354cc7a25d89392cf3b1625417bc775b130c883db2276d86b452575c9483b7b"} Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.230477 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.237009 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.412733 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5b0bee27-5168-4984-9943-c93299d3ac58-util\") pod \"5b0bee27-5168-4984-9943-c93299d3ac58\" (UID: \"5b0bee27-5168-4984-9943-c93299d3ac58\") " Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.413185 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n82z\" (UniqueName: \"kubernetes.io/projected/5b0bee27-5168-4984-9943-c93299d3ac58-kube-api-access-6n82z\") pod \"5b0bee27-5168-4984-9943-c93299d3ac58\" (UID: \"5b0bee27-5168-4984-9943-c93299d3ac58\") " Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.413235 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d0befe17-9e51-457b-86c7-8831cb064339-util\") pod \"d0befe17-9e51-457b-86c7-8831cb064339\" (UID: \"d0befe17-9e51-457b-86c7-8831cb064339\") " Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.413266 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d0befe17-9e51-457b-86c7-8831cb064339-bundle\") pod \"d0befe17-9e51-457b-86c7-8831cb064339\" (UID: \"d0befe17-9e51-457b-86c7-8831cb064339\") " Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.413314 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5b0bee27-5168-4984-9943-c93299d3ac58-bundle\") pod \"5b0bee27-5168-4984-9943-c93299d3ac58\" (UID: \"5b0bee27-5168-4984-9943-c93299d3ac58\") " Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.413343 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njfk4\" (UniqueName: \"kubernetes.io/projected/d0befe17-9e51-457b-86c7-8831cb064339-kube-api-access-njfk4\") pod \"d0befe17-9e51-457b-86c7-8831cb064339\" (UID: \"d0befe17-9e51-457b-86c7-8831cb064339\") " Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.414014 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0befe17-9e51-457b-86c7-8831cb064339-bundle" (OuterVolumeSpecName: "bundle") pod "d0befe17-9e51-457b-86c7-8831cb064339" (UID: "d0befe17-9e51-457b-86c7-8831cb064339"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.414974 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b0bee27-5168-4984-9943-c93299d3ac58-bundle" (OuterVolumeSpecName: "bundle") pod "5b0bee27-5168-4984-9943-c93299d3ac58" (UID: "5b0bee27-5168-4984-9943-c93299d3ac58"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.418700 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b0bee27-5168-4984-9943-c93299d3ac58-kube-api-access-6n82z" (OuterVolumeSpecName: "kube-api-access-6n82z") pod "5b0bee27-5168-4984-9943-c93299d3ac58" (UID: "5b0bee27-5168-4984-9943-c93299d3ac58"). InnerVolumeSpecName "kube-api-access-6n82z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.418763 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0befe17-9e51-457b-86c7-8831cb064339-kube-api-access-njfk4" (OuterVolumeSpecName: "kube-api-access-njfk4") pod "d0befe17-9e51-457b-86c7-8831cb064339" (UID: "d0befe17-9e51-457b-86c7-8831cb064339"). InnerVolumeSpecName "kube-api-access-njfk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.423632 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b0bee27-5168-4984-9943-c93299d3ac58-util" (OuterVolumeSpecName: "util") pod "5b0bee27-5168-4984-9943-c93299d3ac58" (UID: "5b0bee27-5168-4984-9943-c93299d3ac58"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.425257 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0befe17-9e51-457b-86c7-8831cb064339-util" (OuterVolumeSpecName: "util") pod "d0befe17-9e51-457b-86c7-8831cb064339" (UID: "d0befe17-9e51-457b-86c7-8831cb064339"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.514454 4802 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5b0bee27-5168-4984-9943-c93299d3ac58-util\") on node \"crc\" DevicePath \"\"" Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.514484 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n82z\" (UniqueName: \"kubernetes.io/projected/5b0bee27-5168-4984-9943-c93299d3ac58-kube-api-access-6n82z\") on node \"crc\" DevicePath \"\"" Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.514496 4802 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d0befe17-9e51-457b-86c7-8831cb064339-util\") on node \"crc\" DevicePath \"\"" Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.514506 4802 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d0befe17-9e51-457b-86c7-8831cb064339-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.514517 4802 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5b0bee27-5168-4984-9943-c93299d3ac58-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.514526 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njfk4\" (UniqueName: \"kubernetes.io/projected/d0befe17-9e51-457b-86c7-8831cb064339-kube-api-access-njfk4\") on node \"crc\" DevicePath \"\"" Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.966482 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" event={"ID":"5b0bee27-5168-4984-9943-c93299d3ac58","Type":"ContainerDied","Data":"124c9bfe0f4f7bf1e0b076658079cb6ddb7c7599e5c9e4920a8b2f6996d973b3"} Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.966517 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7" Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.966526 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="124c9bfe0f4f7bf1e0b076658079cb6ddb7c7599e5c9e4920a8b2f6996d973b3" Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.968902 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" event={"ID":"d0befe17-9e51-457b-86c7-8831cb064339","Type":"ContainerDied","Data":"89bfc345d4346bde83cd3bef262687e962c6780ef0012c9cd62b7fb1c8c8f32f"} Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.969056 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89bfc345d4346bde83cd3bef262687e962c6780ef0012c9cd62b7fb1c8c8f32f" Nov 25 17:04:46 crc kubenswrapper[4802]: I1125 17:04:46.968999 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr" Nov 25 17:04:50 crc kubenswrapper[4802]: I1125 17:04:50.704291 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/keystone-64bc67949c-jgvwk" Nov 25 17:04:54 crc kubenswrapper[4802]: I1125 17:04:54.969015 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c"] Nov 25 17:04:54 crc kubenswrapper[4802]: E1125 17:04:54.969841 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0bee27-5168-4984-9943-c93299d3ac58" containerName="util" Nov 25 17:04:54 crc kubenswrapper[4802]: I1125 17:04:54.969858 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0bee27-5168-4984-9943-c93299d3ac58" containerName="util" Nov 25 17:04:54 crc kubenswrapper[4802]: E1125 17:04:54.969873 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0bee27-5168-4984-9943-c93299d3ac58" containerName="extract" Nov 25 17:04:54 crc kubenswrapper[4802]: I1125 17:04:54.969882 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0bee27-5168-4984-9943-c93299d3ac58" containerName="extract" Nov 25 17:04:54 crc kubenswrapper[4802]: E1125 17:04:54.969895 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0befe17-9e51-457b-86c7-8831cb064339" containerName="pull" Nov 25 17:04:54 crc kubenswrapper[4802]: I1125 17:04:54.969904 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0befe17-9e51-457b-86c7-8831cb064339" containerName="pull" Nov 25 17:04:54 crc kubenswrapper[4802]: E1125 17:04:54.969928 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0befe17-9e51-457b-86c7-8831cb064339" containerName="extract" Nov 25 17:04:54 crc kubenswrapper[4802]: I1125 17:04:54.969937 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0befe17-9e51-457b-86c7-8831cb064339" containerName="extract" Nov 25 17:04:54 crc kubenswrapper[4802]: E1125 17:04:54.969950 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0befe17-9e51-457b-86c7-8831cb064339" containerName="util" Nov 25 17:04:54 crc kubenswrapper[4802]: I1125 17:04:54.969957 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0befe17-9e51-457b-86c7-8831cb064339" containerName="util" Nov 25 17:04:54 crc kubenswrapper[4802]: E1125 17:04:54.969970 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0bee27-5168-4984-9943-c93299d3ac58" containerName="pull" Nov 25 17:04:54 crc kubenswrapper[4802]: I1125 17:04:54.969978 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0bee27-5168-4984-9943-c93299d3ac58" containerName="pull" Nov 25 17:04:54 crc kubenswrapper[4802]: I1125 17:04:54.970146 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b0bee27-5168-4984-9943-c93299d3ac58" containerName="extract" Nov 25 17:04:54 crc kubenswrapper[4802]: I1125 17:04:54.970161 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0befe17-9e51-457b-86c7-8831cb064339" containerName="extract" Nov 25 17:04:54 crc kubenswrapper[4802]: I1125 17:04:54.970683 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:04:54 crc kubenswrapper[4802]: I1125 17:04:54.973434 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Nov 25 17:04:54 crc kubenswrapper[4802]: I1125 17:04:54.977516 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-nlkzd" Nov 25 17:04:54 crc kubenswrapper[4802]: I1125 17:04:54.985082 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c"] Nov 25 17:04:55 crc kubenswrapper[4802]: I1125 17:04:55.128273 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb-webhook-cert\") pod \"swift-operator-controller-manager-5787d846c-g6b4c\" (UID: \"3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb\") " pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:04:55 crc kubenswrapper[4802]: I1125 17:04:55.128346 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb-apiservice-cert\") pod \"swift-operator-controller-manager-5787d846c-g6b4c\" (UID: \"3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb\") " pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:04:55 crc kubenswrapper[4802]: I1125 17:04:55.128366 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rx765\" (UniqueName: \"kubernetes.io/projected/3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb-kube-api-access-rx765\") pod \"swift-operator-controller-manager-5787d846c-g6b4c\" (UID: \"3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb\") " pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:04:55 crc kubenswrapper[4802]: I1125 17:04:55.230180 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb-webhook-cert\") pod \"swift-operator-controller-manager-5787d846c-g6b4c\" (UID: \"3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb\") " pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:04:55 crc kubenswrapper[4802]: I1125 17:04:55.230234 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb-apiservice-cert\") pod \"swift-operator-controller-manager-5787d846c-g6b4c\" (UID: \"3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb\") " pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:04:55 crc kubenswrapper[4802]: I1125 17:04:55.230253 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rx765\" (UniqueName: \"kubernetes.io/projected/3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb-kube-api-access-rx765\") pod \"swift-operator-controller-manager-5787d846c-g6b4c\" (UID: \"3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb\") " pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:04:55 crc kubenswrapper[4802]: I1125 17:04:55.236381 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb-webhook-cert\") pod \"swift-operator-controller-manager-5787d846c-g6b4c\" (UID: \"3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb\") " pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:04:55 crc kubenswrapper[4802]: I1125 17:04:55.237587 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb-apiservice-cert\") pod \"swift-operator-controller-manager-5787d846c-g6b4c\" (UID: \"3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb\") " pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:04:55 crc kubenswrapper[4802]: I1125 17:04:55.246592 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rx765\" (UniqueName: \"kubernetes.io/projected/3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb-kube-api-access-rx765\") pod \"swift-operator-controller-manager-5787d846c-g6b4c\" (UID: \"3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb\") " pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:04:55 crc kubenswrapper[4802]: I1125 17:04:55.290042 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:04:55 crc kubenswrapper[4802]: I1125 17:04:55.729059 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c"] Nov 25 17:04:56 crc kubenswrapper[4802]: I1125 17:04:56.026442 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" event={"ID":"3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb","Type":"ContainerStarted","Data":"b993716f401dbcee2297e1655d5286360a77b3106d819511ef3899ec29a0225d"} Nov 25 17:04:59 crc kubenswrapper[4802]: I1125 17:04:59.051279 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" event={"ID":"3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb","Type":"ContainerStarted","Data":"342193f4dc464f0cf86d641c2c4e239dbe6d8a0ae54e1e1ef2fb90de1debc0d4"} Nov 25 17:04:59 crc kubenswrapper[4802]: I1125 17:04:59.051916 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:04:59 crc kubenswrapper[4802]: I1125 17:04:59.070473 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" podStartSLOduration=2.001333722 podStartE2EDuration="5.070455566s" podCreationTimestamp="2025-11-25 17:04:54 +0000 UTC" firstStartedPulling="2025-11-25 17:04:55.737179107 +0000 UTC m=+1078.881526283" lastFinishedPulling="2025-11-25 17:04:58.806300941 +0000 UTC m=+1081.950648127" observedRunningTime="2025-11-25 17:04:59.069563002 +0000 UTC m=+1082.213910208" watchObservedRunningTime="2025-11-25 17:04:59.070455566 +0000 UTC m=+1082.214802752" Nov 25 17:05:02 crc kubenswrapper[4802]: I1125 17:05:02.034110 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z"] Nov 25 17:05:02 crc kubenswrapper[4802]: I1125 17:05:02.035578 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:05:02 crc kubenswrapper[4802]: I1125 17:05:02.037591 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Nov 25 17:05:02 crc kubenswrapper[4802]: I1125 17:05:02.038225 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-thlvn" Nov 25 17:05:02 crc kubenswrapper[4802]: I1125 17:05:02.045411 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z"] Nov 25 17:05:02 crc kubenswrapper[4802]: I1125 17:05:02.138164 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rkvx\" (UniqueName: \"kubernetes.io/projected/c328497f-325f-4500-ba48-d1523cd7ee74-kube-api-access-7rkvx\") pod \"horizon-operator-controller-manager-d58bc84f4-fjc7z\" (UID: \"c328497f-325f-4500-ba48-d1523cd7ee74\") " pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:05:02 crc kubenswrapper[4802]: I1125 17:05:02.138236 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c328497f-325f-4500-ba48-d1523cd7ee74-apiservice-cert\") pod \"horizon-operator-controller-manager-d58bc84f4-fjc7z\" (UID: \"c328497f-325f-4500-ba48-d1523cd7ee74\") " pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:05:02 crc kubenswrapper[4802]: I1125 17:05:02.138401 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c328497f-325f-4500-ba48-d1523cd7ee74-webhook-cert\") pod \"horizon-operator-controller-manager-d58bc84f4-fjc7z\" (UID: \"c328497f-325f-4500-ba48-d1523cd7ee74\") " pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:05:02 crc kubenswrapper[4802]: I1125 17:05:02.239621 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c328497f-325f-4500-ba48-d1523cd7ee74-webhook-cert\") pod \"horizon-operator-controller-manager-d58bc84f4-fjc7z\" (UID: \"c328497f-325f-4500-ba48-d1523cd7ee74\") " pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:05:02 crc kubenswrapper[4802]: I1125 17:05:02.239962 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rkvx\" (UniqueName: \"kubernetes.io/projected/c328497f-325f-4500-ba48-d1523cd7ee74-kube-api-access-7rkvx\") pod \"horizon-operator-controller-manager-d58bc84f4-fjc7z\" (UID: \"c328497f-325f-4500-ba48-d1523cd7ee74\") " pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:05:02 crc kubenswrapper[4802]: I1125 17:05:02.239998 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c328497f-325f-4500-ba48-d1523cd7ee74-apiservice-cert\") pod \"horizon-operator-controller-manager-d58bc84f4-fjc7z\" (UID: \"c328497f-325f-4500-ba48-d1523cd7ee74\") " pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:05:02 crc kubenswrapper[4802]: I1125 17:05:02.247073 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c328497f-325f-4500-ba48-d1523cd7ee74-webhook-cert\") pod \"horizon-operator-controller-manager-d58bc84f4-fjc7z\" (UID: \"c328497f-325f-4500-ba48-d1523cd7ee74\") " pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:05:02 crc kubenswrapper[4802]: I1125 17:05:02.247090 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c328497f-325f-4500-ba48-d1523cd7ee74-apiservice-cert\") pod \"horizon-operator-controller-manager-d58bc84f4-fjc7z\" (UID: \"c328497f-325f-4500-ba48-d1523cd7ee74\") " pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:05:02 crc kubenswrapper[4802]: I1125 17:05:02.256352 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rkvx\" (UniqueName: \"kubernetes.io/projected/c328497f-325f-4500-ba48-d1523cd7ee74-kube-api-access-7rkvx\") pod \"horizon-operator-controller-manager-d58bc84f4-fjc7z\" (UID: \"c328497f-325f-4500-ba48-d1523cd7ee74\") " pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:05:02 crc kubenswrapper[4802]: I1125 17:05:02.353902 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:05:02 crc kubenswrapper[4802]: I1125 17:05:02.758141 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z"] Nov 25 17:05:02 crc kubenswrapper[4802]: W1125 17:05:02.767960 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc328497f_325f_4500_ba48_d1523cd7ee74.slice/crio-a3b6654503066ed45811f37c751ed61ad67273274915b38489447793c12db65b WatchSource:0}: Error finding container a3b6654503066ed45811f37c751ed61ad67273274915b38489447793c12db65b: Status 404 returned error can't find the container with id a3b6654503066ed45811f37c751ed61ad67273274915b38489447793c12db65b Nov 25 17:05:03 crc kubenswrapper[4802]: I1125 17:05:03.074752 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" event={"ID":"c328497f-325f-4500-ba48-d1523cd7ee74","Type":"ContainerStarted","Data":"a3b6654503066ed45811f37c751ed61ad67273274915b38489447793c12db65b"} Nov 25 17:05:05 crc kubenswrapper[4802]: I1125 17:05:05.096637 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" event={"ID":"c328497f-325f-4500-ba48-d1523cd7ee74","Type":"ContainerStarted","Data":"9861fdf30e2ca091950f978e9221e687d7cdd4a0c39d27cb97f49db9d42753fa"} Nov 25 17:05:05 crc kubenswrapper[4802]: I1125 17:05:05.097005 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:05:05 crc kubenswrapper[4802]: I1125 17:05:05.125311 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" podStartSLOduration=1.515520489 podStartE2EDuration="3.125291054s" podCreationTimestamp="2025-11-25 17:05:02 +0000 UTC" firstStartedPulling="2025-11-25 17:05:02.770079128 +0000 UTC m=+1085.914426314" lastFinishedPulling="2025-11-25 17:05:04.379849693 +0000 UTC m=+1087.524196879" observedRunningTime="2025-11-25 17:05:05.118507381 +0000 UTC m=+1088.262854567" watchObservedRunningTime="2025-11-25 17:05:05.125291054 +0000 UTC m=+1088.269638240" Nov 25 17:05:05 crc kubenswrapper[4802]: I1125 17:05:05.295241 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:05:10 crc kubenswrapper[4802]: I1125 17:05:10.900108 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-sdnvr"] Nov 25 17:05:10 crc kubenswrapper[4802]: I1125 17:05:10.901463 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-sdnvr" Nov 25 17:05:10 crc kubenswrapper[4802]: I1125 17:05:10.903239 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-9j9ft" Nov 25 17:05:10 crc kubenswrapper[4802]: I1125 17:05:10.950082 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-sdnvr"] Nov 25 17:05:11 crc kubenswrapper[4802]: I1125 17:05:11.086784 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxrc8\" (UniqueName: \"kubernetes.io/projected/04dd5ba3-51fb-4edc-9d85-65fde00950de-kube-api-access-sxrc8\") pod \"glance-operator-index-sdnvr\" (UID: \"04dd5ba3-51fb-4edc-9d85-65fde00950de\") " pod="openstack-operators/glance-operator-index-sdnvr" Nov 25 17:05:11 crc kubenswrapper[4802]: I1125 17:05:11.188015 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxrc8\" (UniqueName: \"kubernetes.io/projected/04dd5ba3-51fb-4edc-9d85-65fde00950de-kube-api-access-sxrc8\") pod \"glance-operator-index-sdnvr\" (UID: \"04dd5ba3-51fb-4edc-9d85-65fde00950de\") " pod="openstack-operators/glance-operator-index-sdnvr" Nov 25 17:05:11 crc kubenswrapper[4802]: I1125 17:05:11.213969 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxrc8\" (UniqueName: \"kubernetes.io/projected/04dd5ba3-51fb-4edc-9d85-65fde00950de-kube-api-access-sxrc8\") pod \"glance-operator-index-sdnvr\" (UID: \"04dd5ba3-51fb-4edc-9d85-65fde00950de\") " pod="openstack-operators/glance-operator-index-sdnvr" Nov 25 17:05:11 crc kubenswrapper[4802]: I1125 17:05:11.222802 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-sdnvr" Nov 25 17:05:11 crc kubenswrapper[4802]: I1125 17:05:11.673078 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-sdnvr"] Nov 25 17:05:12 crc kubenswrapper[4802]: I1125 17:05:12.143021 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-sdnvr" event={"ID":"04dd5ba3-51fb-4edc-9d85-65fde00950de","Type":"ContainerStarted","Data":"c8be67a7fdd77bd1902d5f510b4f7b311644678ac8619d1c1f08c65f1b1e6dd7"} Nov 25 17:05:12 crc kubenswrapper[4802]: I1125 17:05:12.361425 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.335583 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.341475 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.343780 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.344185 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-f9sj6" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.345298 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.345535 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.367483 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.542542 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4a121db3-b69c-484d-9917-f13c0c8ac50f-cache\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.542971 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.543031 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr694\" (UniqueName: \"kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-kube-api-access-cr694\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.543084 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-etc-swift\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.543241 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4a121db3-b69c-484d-9917-f13c0c8ac50f-lock\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.644397 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4a121db3-b69c-484d-9917-f13c0c8ac50f-lock\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.644477 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4a121db3-b69c-484d-9917-f13c0c8ac50f-cache\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.644520 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.644568 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr694\" (UniqueName: \"kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-kube-api-access-cr694\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.644631 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-etc-swift\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:14 crc kubenswrapper[4802]: E1125 17:05:14.644787 4802 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 17:05:14 crc kubenswrapper[4802]: E1125 17:05:14.644801 4802 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 17:05:14 crc kubenswrapper[4802]: E1125 17:05:14.644849 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-etc-swift podName:4a121db3-b69c-484d-9917-f13c0c8ac50f nodeName:}" failed. No retries permitted until 2025-11-25 17:05:15.144831261 +0000 UTC m=+1098.289178447 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-etc-swift") pod "swift-storage-0" (UID: "4a121db3-b69c-484d-9917-f13c0c8ac50f") : configmap "swift-ring-files" not found Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.645630 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4a121db3-b69c-484d-9917-f13c0c8ac50f-lock\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.645841 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4a121db3-b69c-484d-9917-f13c0c8ac50f-cache\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.646632 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.666293 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.670394 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr694\" (UniqueName: \"kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-kube-api-access-cr694\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.835757 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-v7xh9"] Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.837248 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.839003 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.839336 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-config-data" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.839485 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-scripts" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.847593 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-v7xh9"] Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.867103 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-v7xh9"] Nov 25 17:05:14 crc kubenswrapper[4802]: E1125 17:05:14.867818 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[dispersionconf etc-swift kube-api-access-blbvh ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" podUID="9eedb724-1f5d-4e15-b8ec-412bc8d98684" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.892825 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-m4h5c"] Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.893798 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.908824 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-m4h5c"] Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.950986 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9eedb724-1f5d-4e15-b8ec-412bc8d98684-swiftconf\") pod \"swift-ring-rebalance-v7xh9\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.951055 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9eedb724-1f5d-4e15-b8ec-412bc8d98684-scripts\") pod \"swift-ring-rebalance-v7xh9\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.951091 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9eedb724-1f5d-4e15-b8ec-412bc8d98684-dispersionconf\") pod \"swift-ring-rebalance-v7xh9\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.951146 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blbvh\" (UniqueName: \"kubernetes.io/projected/9eedb724-1f5d-4e15-b8ec-412bc8d98684-kube-api-access-blbvh\") pod \"swift-ring-rebalance-v7xh9\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.951178 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9eedb724-1f5d-4e15-b8ec-412bc8d98684-etc-swift\") pod \"swift-ring-rebalance-v7xh9\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:14 crc kubenswrapper[4802]: I1125 17:05:14.951288 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9eedb724-1f5d-4e15-b8ec-412bc8d98684-ring-data-devices\") pod \"swift-ring-rebalance-v7xh9\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.052396 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s66m\" (UniqueName: \"kubernetes.io/projected/305295e6-42a5-41a9-ad55-17945c616b87-kube-api-access-8s66m\") pod \"swift-ring-rebalance-m4h5c\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.052457 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9eedb724-1f5d-4e15-b8ec-412bc8d98684-ring-data-devices\") pod \"swift-ring-rebalance-v7xh9\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.052521 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/305295e6-42a5-41a9-ad55-17945c616b87-etc-swift\") pod \"swift-ring-rebalance-m4h5c\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.052544 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/305295e6-42a5-41a9-ad55-17945c616b87-ring-data-devices\") pod \"swift-ring-rebalance-m4h5c\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.052575 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/305295e6-42a5-41a9-ad55-17945c616b87-scripts\") pod \"swift-ring-rebalance-m4h5c\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.052600 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9eedb724-1f5d-4e15-b8ec-412bc8d98684-swiftconf\") pod \"swift-ring-rebalance-v7xh9\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.052690 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9eedb724-1f5d-4e15-b8ec-412bc8d98684-scripts\") pod \"swift-ring-rebalance-v7xh9\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.052720 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9eedb724-1f5d-4e15-b8ec-412bc8d98684-dispersionconf\") pod \"swift-ring-rebalance-v7xh9\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.052748 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/305295e6-42a5-41a9-ad55-17945c616b87-dispersionconf\") pod \"swift-ring-rebalance-m4h5c\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.052773 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blbvh\" (UniqueName: \"kubernetes.io/projected/9eedb724-1f5d-4e15-b8ec-412bc8d98684-kube-api-access-blbvh\") pod \"swift-ring-rebalance-v7xh9\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.052803 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9eedb724-1f5d-4e15-b8ec-412bc8d98684-etc-swift\") pod \"swift-ring-rebalance-v7xh9\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.053294 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9eedb724-1f5d-4e15-b8ec-412bc8d98684-etc-swift\") pod \"swift-ring-rebalance-v7xh9\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.053331 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9eedb724-1f5d-4e15-b8ec-412bc8d98684-ring-data-devices\") pod \"swift-ring-rebalance-v7xh9\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.053344 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/305295e6-42a5-41a9-ad55-17945c616b87-swiftconf\") pod \"swift-ring-rebalance-m4h5c\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.053529 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9eedb724-1f5d-4e15-b8ec-412bc8d98684-scripts\") pod \"swift-ring-rebalance-v7xh9\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.057490 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9eedb724-1f5d-4e15-b8ec-412bc8d98684-swiftconf\") pod \"swift-ring-rebalance-v7xh9\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.057873 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9eedb724-1f5d-4e15-b8ec-412bc8d98684-dispersionconf\") pod \"swift-ring-rebalance-v7xh9\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.077090 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blbvh\" (UniqueName: \"kubernetes.io/projected/9eedb724-1f5d-4e15-b8ec-412bc8d98684-kube-api-access-blbvh\") pod \"swift-ring-rebalance-v7xh9\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.154806 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/305295e6-42a5-41a9-ad55-17945c616b87-dispersionconf\") pod \"swift-ring-rebalance-m4h5c\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.154870 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-etc-swift\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.154909 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/305295e6-42a5-41a9-ad55-17945c616b87-swiftconf\") pod \"swift-ring-rebalance-m4h5c\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.154941 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s66m\" (UniqueName: \"kubernetes.io/projected/305295e6-42a5-41a9-ad55-17945c616b87-kube-api-access-8s66m\") pod \"swift-ring-rebalance-m4h5c\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.154977 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/305295e6-42a5-41a9-ad55-17945c616b87-etc-swift\") pod \"swift-ring-rebalance-m4h5c\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.154992 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/305295e6-42a5-41a9-ad55-17945c616b87-ring-data-devices\") pod \"swift-ring-rebalance-m4h5c\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.155011 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/305295e6-42a5-41a9-ad55-17945c616b87-scripts\") pod \"swift-ring-rebalance-m4h5c\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.155883 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/305295e6-42a5-41a9-ad55-17945c616b87-scripts\") pod \"swift-ring-rebalance-m4h5c\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: E1125 17:05:15.156879 4802 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 17:05:15 crc kubenswrapper[4802]: E1125 17:05:15.156909 4802 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 17:05:15 crc kubenswrapper[4802]: E1125 17:05:15.156958 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-etc-swift podName:4a121db3-b69c-484d-9917-f13c0c8ac50f nodeName:}" failed. No retries permitted until 2025-11-25 17:05:16.156937933 +0000 UTC m=+1099.301285159 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-etc-swift") pod "swift-storage-0" (UID: "4a121db3-b69c-484d-9917-f13c0c8ac50f") : configmap "swift-ring-files" not found Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.157336 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/305295e6-42a5-41a9-ad55-17945c616b87-etc-swift\") pod \"swift-ring-rebalance-m4h5c\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.158894 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/305295e6-42a5-41a9-ad55-17945c616b87-ring-data-devices\") pod \"swift-ring-rebalance-m4h5c\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.165252 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/305295e6-42a5-41a9-ad55-17945c616b87-dispersionconf\") pod \"swift-ring-rebalance-m4h5c\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.174420 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/305295e6-42a5-41a9-ad55-17945c616b87-swiftconf\") pod \"swift-ring-rebalance-m4h5c\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.177488 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s66m\" (UniqueName: \"kubernetes.io/projected/305295e6-42a5-41a9-ad55-17945c616b87-kube-api-access-8s66m\") pod \"swift-ring-rebalance-m4h5c\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.182961 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.192775 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.210966 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.271294 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb"] Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.272543 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.282253 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb"] Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.357785 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9eedb724-1f5d-4e15-b8ec-412bc8d98684-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9eedb724-1f5d-4e15-b8ec-412bc8d98684" (UID: "9eedb724-1f5d-4e15-b8ec-412bc8d98684"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.357363 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9eedb724-1f5d-4e15-b8ec-412bc8d98684-etc-swift\") pod \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.357993 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9eedb724-1f5d-4e15-b8ec-412bc8d98684-ring-data-devices\") pod \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.358412 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9eedb724-1f5d-4e15-b8ec-412bc8d98684-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "9eedb724-1f5d-4e15-b8ec-412bc8d98684" (UID: "9eedb724-1f5d-4e15-b8ec-412bc8d98684"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.358480 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9eedb724-1f5d-4e15-b8ec-412bc8d98684-scripts\") pod \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.358867 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9eedb724-1f5d-4e15-b8ec-412bc8d98684-scripts" (OuterVolumeSpecName: "scripts") pod "9eedb724-1f5d-4e15-b8ec-412bc8d98684" (UID: "9eedb724-1f5d-4e15-b8ec-412bc8d98684"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.359032 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9eedb724-1f5d-4e15-b8ec-412bc8d98684-dispersionconf\") pod \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.359072 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blbvh\" (UniqueName: \"kubernetes.io/projected/9eedb724-1f5d-4e15-b8ec-412bc8d98684-kube-api-access-blbvh\") pod \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.359105 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9eedb724-1f5d-4e15-b8ec-412bc8d98684-swiftconf\") pod \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\" (UID: \"9eedb724-1f5d-4e15-b8ec-412bc8d98684\") " Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.359583 4802 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9eedb724-1f5d-4e15-b8ec-412bc8d98684-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.359596 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9eedb724-1f5d-4e15-b8ec-412bc8d98684-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.359605 4802 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9eedb724-1f5d-4e15-b8ec-412bc8d98684-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.364378 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eedb724-1f5d-4e15-b8ec-412bc8d98684-kube-api-access-blbvh" (OuterVolumeSpecName: "kube-api-access-blbvh") pod "9eedb724-1f5d-4e15-b8ec-412bc8d98684" (UID: "9eedb724-1f5d-4e15-b8ec-412bc8d98684"). InnerVolumeSpecName "kube-api-access-blbvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.364508 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9eedb724-1f5d-4e15-b8ec-412bc8d98684-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "9eedb724-1f5d-4e15-b8ec-412bc8d98684" (UID: "9eedb724-1f5d-4e15-b8ec-412bc8d98684"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.376735 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9eedb724-1f5d-4e15-b8ec-412bc8d98684-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "9eedb724-1f5d-4e15-b8ec-412bc8d98684" (UID: "9eedb724-1f5d-4e15-b8ec-412bc8d98684"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.461042 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b82f4893-deaf-4760-8448-30dbc4c5e100-config-data\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.461161 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b82f4893-deaf-4760-8448-30dbc4c5e100-run-httpd\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.461453 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-etc-swift\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.461634 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk2jn\" (UniqueName: \"kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-kube-api-access-qk2jn\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.461740 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b82f4893-deaf-4760-8448-30dbc4c5e100-log-httpd\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.461993 4802 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9eedb724-1f5d-4e15-b8ec-412bc8d98684-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.462013 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blbvh\" (UniqueName: \"kubernetes.io/projected/9eedb724-1f5d-4e15-b8ec-412bc8d98684-kube-api-access-blbvh\") on node \"crc\" DevicePath \"\"" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.462026 4802 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9eedb724-1f5d-4e15-b8ec-412bc8d98684-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.563171 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b82f4893-deaf-4760-8448-30dbc4c5e100-config-data\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.563319 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b82f4893-deaf-4760-8448-30dbc4c5e100-run-httpd\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.563462 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-etc-swift\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:15 crc kubenswrapper[4802]: E1125 17:05:15.563714 4802 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 17:05:15 crc kubenswrapper[4802]: E1125 17:05:15.563748 4802 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb: configmap "swift-ring-files" not found Nov 25 17:05:15 crc kubenswrapper[4802]: E1125 17:05:15.563828 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-etc-swift podName:b82f4893-deaf-4760-8448-30dbc4c5e100 nodeName:}" failed. No retries permitted until 2025-11-25 17:05:16.063803654 +0000 UTC m=+1099.208150910 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-etc-swift") pod "swift-proxy-6bd58cfcf7-4q4kb" (UID: "b82f4893-deaf-4760-8448-30dbc4c5e100") : configmap "swift-ring-files" not found Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.563974 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b82f4893-deaf-4760-8448-30dbc4c5e100-run-httpd\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.563512 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk2jn\" (UniqueName: \"kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-kube-api-access-qk2jn\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.564057 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b82f4893-deaf-4760-8448-30dbc4c5e100-log-httpd\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.564427 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b82f4893-deaf-4760-8448-30dbc4c5e100-log-httpd\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.567097 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b82f4893-deaf-4760-8448-30dbc4c5e100-config-data\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:15 crc kubenswrapper[4802]: I1125 17:05:15.585732 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk2jn\" (UniqueName: \"kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-kube-api-access-qk2jn\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:16 crc kubenswrapper[4802]: I1125 17:05:16.076088 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-etc-swift\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:16 crc kubenswrapper[4802]: E1125 17:05:16.076252 4802 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 17:05:16 crc kubenswrapper[4802]: E1125 17:05:16.076499 4802 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb: configmap "swift-ring-files" not found Nov 25 17:05:16 crc kubenswrapper[4802]: E1125 17:05:16.076549 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-etc-swift podName:b82f4893-deaf-4760-8448-30dbc4c5e100 nodeName:}" failed. No retries permitted until 2025-11-25 17:05:17.076531772 +0000 UTC m=+1100.220878948 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-etc-swift") pod "swift-proxy-6bd58cfcf7-4q4kb" (UID: "b82f4893-deaf-4760-8448-30dbc4c5e100") : configmap "swift-ring-files" not found Nov 25 17:05:16 crc kubenswrapper[4802]: I1125 17:05:16.184058 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-etc-swift\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:16 crc kubenswrapper[4802]: E1125 17:05:16.184239 4802 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 17:05:16 crc kubenswrapper[4802]: E1125 17:05:16.184264 4802 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 17:05:16 crc kubenswrapper[4802]: E1125 17:05:16.184323 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-etc-swift podName:4a121db3-b69c-484d-9917-f13c0c8ac50f nodeName:}" failed. No retries permitted until 2025-11-25 17:05:18.184300541 +0000 UTC m=+1101.328647727 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-etc-swift") pod "swift-storage-0" (UID: "4a121db3-b69c-484d-9917-f13c0c8ac50f") : configmap "swift-ring-files" not found Nov 25 17:05:16 crc kubenswrapper[4802]: I1125 17:05:16.192872 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-v7xh9" Nov 25 17:05:16 crc kubenswrapper[4802]: I1125 17:05:16.238185 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-v7xh9"] Nov 25 17:05:16 crc kubenswrapper[4802]: I1125 17:05:16.250256 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-v7xh9"] Nov 25 17:05:16 crc kubenswrapper[4802]: I1125 17:05:16.506898 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-m4h5c"] Nov 25 17:05:17 crc kubenswrapper[4802]: I1125 17:05:17.101221 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-etc-swift\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:17 crc kubenswrapper[4802]: E1125 17:05:17.101447 4802 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 17:05:17 crc kubenswrapper[4802]: E1125 17:05:17.101474 4802 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb: configmap "swift-ring-files" not found Nov 25 17:05:17 crc kubenswrapper[4802]: E1125 17:05:17.101540 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-etc-swift podName:b82f4893-deaf-4760-8448-30dbc4c5e100 nodeName:}" failed. No retries permitted until 2025-11-25 17:05:19.101517786 +0000 UTC m=+1102.245864972 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-etc-swift") pod "swift-proxy-6bd58cfcf7-4q4kb" (UID: "b82f4893-deaf-4760-8448-30dbc4c5e100") : configmap "swift-ring-files" not found Nov 25 17:05:17 crc kubenswrapper[4802]: I1125 17:05:17.201658 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" event={"ID":"305295e6-42a5-41a9-ad55-17945c616b87","Type":"ContainerStarted","Data":"abbfdf92ca319950f54cf2ff35a6a27b370bd6e33a3a2fd888ef55bf0e37b0c2"} Nov 25 17:05:17 crc kubenswrapper[4802]: I1125 17:05:17.514751 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9eedb724-1f5d-4e15-b8ec-412bc8d98684" path="/var/lib/kubelet/pods/9eedb724-1f5d-4e15-b8ec-412bc8d98684/volumes" Nov 25 17:05:18 crc kubenswrapper[4802]: I1125 17:05:18.218290 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-etc-swift\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:18 crc kubenswrapper[4802]: E1125 17:05:18.218476 4802 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 17:05:18 crc kubenswrapper[4802]: E1125 17:05:18.218890 4802 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 17:05:18 crc kubenswrapper[4802]: E1125 17:05:18.219010 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-etc-swift podName:4a121db3-b69c-484d-9917-f13c0c8ac50f nodeName:}" failed. No retries permitted until 2025-11-25 17:05:22.218983174 +0000 UTC m=+1105.363330360 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-etc-swift") pod "swift-storage-0" (UID: "4a121db3-b69c-484d-9917-f13c0c8ac50f") : configmap "swift-ring-files" not found Nov 25 17:05:19 crc kubenswrapper[4802]: I1125 17:05:19.136393 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-etc-swift\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:19 crc kubenswrapper[4802]: E1125 17:05:19.136670 4802 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 17:05:19 crc kubenswrapper[4802]: E1125 17:05:19.136879 4802 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb: configmap "swift-ring-files" not found Nov 25 17:05:19 crc kubenswrapper[4802]: E1125 17:05:19.137013 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-etc-swift podName:b82f4893-deaf-4760-8448-30dbc4c5e100 nodeName:}" failed. No retries permitted until 2025-11-25 17:05:23.136981061 +0000 UTC m=+1106.281328237 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-etc-swift") pod "swift-proxy-6bd58cfcf7-4q4kb" (UID: "b82f4893-deaf-4760-8448-30dbc4c5e100") : configmap "swift-ring-files" not found Nov 25 17:05:19 crc kubenswrapper[4802]: I1125 17:05:19.220091 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-sdnvr" event={"ID":"04dd5ba3-51fb-4edc-9d85-65fde00950de","Type":"ContainerStarted","Data":"30cd71bf7b85ecdbede7e876b5c152a29f3843f57cbd7f3d3e40aa8be354795c"} Nov 25 17:05:19 crc kubenswrapper[4802]: I1125 17:05:19.238924 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-sdnvr" podStartSLOduration=2.714137374 podStartE2EDuration="9.238901731s" podCreationTimestamp="2025-11-25 17:05:10 +0000 UTC" firstStartedPulling="2025-11-25 17:05:11.684968287 +0000 UTC m=+1094.829315473" lastFinishedPulling="2025-11-25 17:05:18.209732654 +0000 UTC m=+1101.354079830" observedRunningTime="2025-11-25 17:05:19.236966118 +0000 UTC m=+1102.381313324" watchObservedRunningTime="2025-11-25 17:05:19.238901731 +0000 UTC m=+1102.383248907" Nov 25 17:05:21 crc kubenswrapper[4802]: I1125 17:05:21.222965 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/glance-operator-index-sdnvr" Nov 25 17:05:21 crc kubenswrapper[4802]: I1125 17:05:21.223571 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-index-sdnvr" Nov 25 17:05:21 crc kubenswrapper[4802]: I1125 17:05:21.256659 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" event={"ID":"305295e6-42a5-41a9-ad55-17945c616b87","Type":"ContainerStarted","Data":"f9437400b6afbc9e31f5deb4737adad8e5717f87108625272cca4e6a39965874"} Nov 25 17:05:21 crc kubenswrapper[4802]: I1125 17:05:21.290874 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/glance-operator-index-sdnvr" Nov 25 17:05:21 crc kubenswrapper[4802]: I1125 17:05:21.292508 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" podStartSLOduration=2.786328168 podStartE2EDuration="7.292495956s" podCreationTimestamp="2025-11-25 17:05:14 +0000 UTC" firstStartedPulling="2025-11-25 17:05:16.514469605 +0000 UTC m=+1099.658816781" lastFinishedPulling="2025-11-25 17:05:21.020637383 +0000 UTC m=+1104.164984569" observedRunningTime="2025-11-25 17:05:21.281987912 +0000 UTC m=+1104.426335098" watchObservedRunningTime="2025-11-25 17:05:21.292495956 +0000 UTC m=+1104.436843142" Nov 25 17:05:22 crc kubenswrapper[4802]: I1125 17:05:22.304056 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-etc-swift\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:22 crc kubenswrapper[4802]: E1125 17:05:22.304441 4802 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 17:05:22 crc kubenswrapper[4802]: E1125 17:05:22.304497 4802 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 17:05:22 crc kubenswrapper[4802]: E1125 17:05:22.304615 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-etc-swift podName:4a121db3-b69c-484d-9917-f13c0c8ac50f nodeName:}" failed. No retries permitted until 2025-11-25 17:05:30.304580443 +0000 UTC m=+1113.448927629 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-etc-swift") pod "swift-storage-0" (UID: "4a121db3-b69c-484d-9917-f13c0c8ac50f") : configmap "swift-ring-files" not found Nov 25 17:05:23 crc kubenswrapper[4802]: I1125 17:05:23.218962 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-etc-swift\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:23 crc kubenswrapper[4802]: E1125 17:05:23.219161 4802 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 17:05:23 crc kubenswrapper[4802]: E1125 17:05:23.219484 4802 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb: configmap "swift-ring-files" not found Nov 25 17:05:23 crc kubenswrapper[4802]: E1125 17:05:23.219614 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-etc-swift podName:b82f4893-deaf-4760-8448-30dbc4c5e100 nodeName:}" failed. No retries permitted until 2025-11-25 17:05:31.219576342 +0000 UTC m=+1114.363923538 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-etc-swift") pod "swift-proxy-6bd58cfcf7-4q4kb" (UID: "b82f4893-deaf-4760-8448-30dbc4c5e100") : configmap "swift-ring-files" not found Nov 25 17:05:24 crc kubenswrapper[4802]: I1125 17:05:24.248650 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 17:05:24 crc kubenswrapper[4802]: I1125 17:05:24.248746 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 17:05:29 crc kubenswrapper[4802]: I1125 17:05:29.313633 4802 generic.go:334] "Generic (PLEG): container finished" podID="305295e6-42a5-41a9-ad55-17945c616b87" containerID="f9437400b6afbc9e31f5deb4737adad8e5717f87108625272cca4e6a39965874" exitCode=0 Nov 25 17:05:29 crc kubenswrapper[4802]: I1125 17:05:29.313747 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" event={"ID":"305295e6-42a5-41a9-ad55-17945c616b87","Type":"ContainerDied","Data":"f9437400b6afbc9e31f5deb4737adad8e5717f87108625272cca4e6a39965874"} Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.381736 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-etc-swift\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.390887 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4a121db3-b69c-484d-9917-f13c0c8ac50f-etc-swift\") pod \"swift-storage-0\" (UID: \"4a121db3-b69c-484d-9917-f13c0c8ac50f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.571572 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.639069 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.686630 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/305295e6-42a5-41a9-ad55-17945c616b87-swiftconf\") pod \"305295e6-42a5-41a9-ad55-17945c616b87\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.687319 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/305295e6-42a5-41a9-ad55-17945c616b87-etc-swift\") pod \"305295e6-42a5-41a9-ad55-17945c616b87\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.687374 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8s66m\" (UniqueName: \"kubernetes.io/projected/305295e6-42a5-41a9-ad55-17945c616b87-kube-api-access-8s66m\") pod \"305295e6-42a5-41a9-ad55-17945c616b87\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.687438 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/305295e6-42a5-41a9-ad55-17945c616b87-dispersionconf\") pod \"305295e6-42a5-41a9-ad55-17945c616b87\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.687512 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/305295e6-42a5-41a9-ad55-17945c616b87-ring-data-devices\") pod \"305295e6-42a5-41a9-ad55-17945c616b87\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.687555 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/305295e6-42a5-41a9-ad55-17945c616b87-scripts\") pod \"305295e6-42a5-41a9-ad55-17945c616b87\" (UID: \"305295e6-42a5-41a9-ad55-17945c616b87\") " Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.689081 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/305295e6-42a5-41a9-ad55-17945c616b87-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "305295e6-42a5-41a9-ad55-17945c616b87" (UID: "305295e6-42a5-41a9-ad55-17945c616b87"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.689063 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/305295e6-42a5-41a9-ad55-17945c616b87-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "305295e6-42a5-41a9-ad55-17945c616b87" (UID: "305295e6-42a5-41a9-ad55-17945c616b87"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.708474 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/305295e6-42a5-41a9-ad55-17945c616b87-kube-api-access-8s66m" (OuterVolumeSpecName: "kube-api-access-8s66m") pod "305295e6-42a5-41a9-ad55-17945c616b87" (UID: "305295e6-42a5-41a9-ad55-17945c616b87"). InnerVolumeSpecName "kube-api-access-8s66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.720762 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/305295e6-42a5-41a9-ad55-17945c616b87-scripts" (OuterVolumeSpecName: "scripts") pod "305295e6-42a5-41a9-ad55-17945c616b87" (UID: "305295e6-42a5-41a9-ad55-17945c616b87"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.733342 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/305295e6-42a5-41a9-ad55-17945c616b87-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "305295e6-42a5-41a9-ad55-17945c616b87" (UID: "305295e6-42a5-41a9-ad55-17945c616b87"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.735809 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/305295e6-42a5-41a9-ad55-17945c616b87-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "305295e6-42a5-41a9-ad55-17945c616b87" (UID: "305295e6-42a5-41a9-ad55-17945c616b87"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.789211 4802 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/305295e6-42a5-41a9-ad55-17945c616b87-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.789256 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8s66m\" (UniqueName: \"kubernetes.io/projected/305295e6-42a5-41a9-ad55-17945c616b87-kube-api-access-8s66m\") on node \"crc\" DevicePath \"\"" Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.789267 4802 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/305295e6-42a5-41a9-ad55-17945c616b87-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.789276 4802 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/305295e6-42a5-41a9-ad55-17945c616b87-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.789286 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/305295e6-42a5-41a9-ad55-17945c616b87-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.789294 4802 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/305295e6-42a5-41a9-ad55-17945c616b87-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 25 17:05:30 crc kubenswrapper[4802]: I1125 17:05:30.820416 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 25 17:05:30 crc kubenswrapper[4802]: W1125 17:05:30.824044 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a121db3_b69c_484d_9917_f13c0c8ac50f.slice/crio-3cfcf8fa01c53da6fcbbdd38ebdbca2165e584b124782479113e4253506867de WatchSource:0}: Error finding container 3cfcf8fa01c53da6fcbbdd38ebdbca2165e584b124782479113e4253506867de: Status 404 returned error can't find the container with id 3cfcf8fa01c53da6fcbbdd38ebdbca2165e584b124782479113e4253506867de Nov 25 17:05:31 crc kubenswrapper[4802]: I1125 17:05:31.266692 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-index-sdnvr" Nov 25 17:05:31 crc kubenswrapper[4802]: I1125 17:05:31.297356 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-etc-swift\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:31 crc kubenswrapper[4802]: I1125 17:05:31.303598 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b82f4893-deaf-4760-8448-30dbc4c5e100-etc-swift\") pod \"swift-proxy-6bd58cfcf7-4q4kb\" (UID: \"b82f4893-deaf-4760-8448-30dbc4c5e100\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:31 crc kubenswrapper[4802]: I1125 17:05:31.329108 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" Nov 25 17:05:31 crc kubenswrapper[4802]: I1125 17:05:31.329555 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-m4h5c" event={"ID":"305295e6-42a5-41a9-ad55-17945c616b87","Type":"ContainerDied","Data":"abbfdf92ca319950f54cf2ff35a6a27b370bd6e33a3a2fd888ef55bf0e37b0c2"} Nov 25 17:05:31 crc kubenswrapper[4802]: I1125 17:05:31.329810 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abbfdf92ca319950f54cf2ff35a6a27b370bd6e33a3a2fd888ef55bf0e37b0c2" Nov 25 17:05:31 crc kubenswrapper[4802]: I1125 17:05:31.331036 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"4a121db3-b69c-484d-9917-f13c0c8ac50f","Type":"ContainerStarted","Data":"3cfcf8fa01c53da6fcbbdd38ebdbca2165e584b124782479113e4253506867de"} Nov 25 17:05:31 crc kubenswrapper[4802]: I1125 17:05:31.504115 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:31 crc kubenswrapper[4802]: I1125 17:05:31.748635 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb"] Nov 25 17:05:32 crc kubenswrapper[4802]: W1125 17:05:32.506949 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb82f4893_deaf_4760_8448_30dbc4c5e100.slice/crio-93100c7512db58afb1c7a6325814ccd655bac1928c30a1bdc134939e898d731d WatchSource:0}: Error finding container 93100c7512db58afb1c7a6325814ccd655bac1928c30a1bdc134939e898d731d: Status 404 returned error can't find the container with id 93100c7512db58afb1c7a6325814ccd655bac1928c30a1bdc134939e898d731d Nov 25 17:05:33 crc kubenswrapper[4802]: I1125 17:05:33.349885 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"4a121db3-b69c-484d-9917-f13c0c8ac50f","Type":"ContainerStarted","Data":"53d25a970a4dabfba923149b2c52fcfaac63e0ce37f1037452b78a859a5402a8"} Nov 25 17:05:33 crc kubenswrapper[4802]: I1125 17:05:33.350209 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"4a121db3-b69c-484d-9917-f13c0c8ac50f","Type":"ContainerStarted","Data":"ac50c5ccdf48887279f8ed9abfc048fe323ba72f0f7c0ab37dac04776d44f18d"} Nov 25 17:05:33 crc kubenswrapper[4802]: I1125 17:05:33.350223 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"4a121db3-b69c-484d-9917-f13c0c8ac50f","Type":"ContainerStarted","Data":"da11f12c1b8cb3315dfabbe6291af13b368cc2508be3390d54458e2a42aafb16"} Nov 25 17:05:33 crc kubenswrapper[4802]: I1125 17:05:33.352423 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" event={"ID":"b82f4893-deaf-4760-8448-30dbc4c5e100","Type":"ContainerStarted","Data":"c0851c46ffbe605dc09cf73589c3bc132097ca1ed65a9fb4d851e7cc04f928ae"} Nov 25 17:05:33 crc kubenswrapper[4802]: I1125 17:05:33.352457 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" event={"ID":"b82f4893-deaf-4760-8448-30dbc4c5e100","Type":"ContainerStarted","Data":"11c48b206519c6a18c54de9349aba12f34d512ce655e88357cead32e7d8b080a"} Nov 25 17:05:33 crc kubenswrapper[4802]: I1125 17:05:33.352470 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" event={"ID":"b82f4893-deaf-4760-8448-30dbc4c5e100","Type":"ContainerStarted","Data":"93100c7512db58afb1c7a6325814ccd655bac1928c30a1bdc134939e898d731d"} Nov 25 17:05:33 crc kubenswrapper[4802]: I1125 17:05:33.352653 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:33 crc kubenswrapper[4802]: I1125 17:05:33.373710 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" podStartSLOduration=18.373679811 podStartE2EDuration="18.373679811s" podCreationTimestamp="2025-11-25 17:05:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:05:33.371504712 +0000 UTC m=+1116.515851898" watchObservedRunningTime="2025-11-25 17:05:33.373679811 +0000 UTC m=+1116.518026997" Nov 25 17:05:34 crc kubenswrapper[4802]: I1125 17:05:34.364387 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"4a121db3-b69c-484d-9917-f13c0c8ac50f","Type":"ContainerStarted","Data":"09261a17bfa4b8537a6e42fd78b15755da26bfaa1af4471570d0ed23dd198b0d"} Nov 25 17:05:34 crc kubenswrapper[4802]: I1125 17:05:34.364716 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:35 crc kubenswrapper[4802]: I1125 17:05:35.379004 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"4a121db3-b69c-484d-9917-f13c0c8ac50f","Type":"ContainerStarted","Data":"de7bff78d93b70ea3a585eaa1f295a1ff1fb56d7c427e9b67c41665c21bad00f"} Nov 25 17:05:35 crc kubenswrapper[4802]: I1125 17:05:35.379574 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"4a121db3-b69c-484d-9917-f13c0c8ac50f","Type":"ContainerStarted","Data":"86189df9739ffdccbb8a1461b5dc7e419f5f1c2425cae4ca5cf4cca918bd8f44"} Nov 25 17:05:35 crc kubenswrapper[4802]: I1125 17:05:35.379594 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"4a121db3-b69c-484d-9917-f13c0c8ac50f","Type":"ContainerStarted","Data":"481273f5b67e04a681953e016a051b1642b7e8d38d15c2485165c516b5b60c3c"} Nov 25 17:05:35 crc kubenswrapper[4802]: I1125 17:05:35.379605 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"4a121db3-b69c-484d-9917-f13c0c8ac50f","Type":"ContainerStarted","Data":"0fa7266665bc4dc87f3401c8871b65a2bf341335086f46e59315c0131901ad65"} Nov 25 17:05:37 crc kubenswrapper[4802]: I1125 17:05:37.398901 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"4a121db3-b69c-484d-9917-f13c0c8ac50f","Type":"ContainerStarted","Data":"22659f05c2234339d9853200fcbf823c088877f185c8d9ec86c001304fef3a51"} Nov 25 17:05:37 crc kubenswrapper[4802]: I1125 17:05:37.400164 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"4a121db3-b69c-484d-9917-f13c0c8ac50f","Type":"ContainerStarted","Data":"1b44824d3be88d88d2f91c55e0cd4bcf25fff28b4bc8515587e6b85f20fe2043"} Nov 25 17:05:37 crc kubenswrapper[4802]: I1125 17:05:37.400261 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"4a121db3-b69c-484d-9917-f13c0c8ac50f","Type":"ContainerStarted","Data":"b6021da1dbaf04a937b13ab0a7dad06a42014a0a247589c813b60e051dee8473"} Nov 25 17:05:37 crc kubenswrapper[4802]: I1125 17:05:37.400375 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"4a121db3-b69c-484d-9917-f13c0c8ac50f","Type":"ContainerStarted","Data":"116fc49b4df5a033ebc5dc9513f881011a936e72db9c15e1299eeb2656f01ff4"} Nov 25 17:05:37 crc kubenswrapper[4802]: I1125 17:05:37.400454 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"4a121db3-b69c-484d-9917-f13c0c8ac50f","Type":"ContainerStarted","Data":"fbca8405c77f22648f14d706bc05db726c7b2d5333e245feb1ace505321d8ff1"} Nov 25 17:05:37 crc kubenswrapper[4802]: I1125 17:05:37.400541 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"4a121db3-b69c-484d-9917-f13c0c8ac50f","Type":"ContainerStarted","Data":"71f84968bf29844875e1f09a754f0eb4cf307e978d34f00cb29249ece8d3a88d"} Nov 25 17:05:38 crc kubenswrapper[4802]: I1125 17:05:38.416206 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"4a121db3-b69c-484d-9917-f13c0c8ac50f","Type":"ContainerStarted","Data":"34201051b6daa7916098804080fdd631facdc35d67d47212d224e5591c9b121f"} Nov 25 17:05:38 crc kubenswrapper[4802]: I1125 17:05:38.460691 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-storage-0" podStartSLOduration=19.931742346 podStartE2EDuration="25.460669024s" podCreationTimestamp="2025-11-25 17:05:13 +0000 UTC" firstStartedPulling="2025-11-25 17:05:30.82667499 +0000 UTC m=+1113.971022186" lastFinishedPulling="2025-11-25 17:05:36.355601678 +0000 UTC m=+1119.499948864" observedRunningTime="2025-11-25 17:05:38.455520854 +0000 UTC m=+1121.599868050" watchObservedRunningTime="2025-11-25 17:05:38.460669024 +0000 UTC m=+1121.605016210" Nov 25 17:05:41 crc kubenswrapper[4802]: I1125 17:05:41.514496 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:41 crc kubenswrapper[4802]: I1125 17:05:41.515596 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-4q4kb" Nov 25 17:05:42 crc kubenswrapper[4802]: I1125 17:05:42.939894 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k"] Nov 25 17:05:42 crc kubenswrapper[4802]: E1125 17:05:42.940438 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="305295e6-42a5-41a9-ad55-17945c616b87" containerName="swift-ring-rebalance" Nov 25 17:05:42 crc kubenswrapper[4802]: I1125 17:05:42.940451 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="305295e6-42a5-41a9-ad55-17945c616b87" containerName="swift-ring-rebalance" Nov 25 17:05:42 crc kubenswrapper[4802]: I1125 17:05:42.940587 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="305295e6-42a5-41a9-ad55-17945c616b87" containerName="swift-ring-rebalance" Nov 25 17:05:42 crc kubenswrapper[4802]: I1125 17:05:42.941531 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k" Nov 25 17:05:42 crc kubenswrapper[4802]: I1125 17:05:42.944174 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-t7662" Nov 25 17:05:42 crc kubenswrapper[4802]: I1125 17:05:42.948840 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k"] Nov 25 17:05:42 crc kubenswrapper[4802]: I1125 17:05:42.997453 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9ad4315c-9d47-485c-b656-4ba2e41803f6-util\") pod \"19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k\" (UID: \"9ad4315c-9d47-485c-b656-4ba2e41803f6\") " pod="openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k" Nov 25 17:05:42 crc kubenswrapper[4802]: I1125 17:05:42.997559 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7qst\" (UniqueName: \"kubernetes.io/projected/9ad4315c-9d47-485c-b656-4ba2e41803f6-kube-api-access-j7qst\") pod \"19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k\" (UID: \"9ad4315c-9d47-485c-b656-4ba2e41803f6\") " pod="openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k" Nov 25 17:05:42 crc kubenswrapper[4802]: I1125 17:05:42.997704 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9ad4315c-9d47-485c-b656-4ba2e41803f6-bundle\") pod \"19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k\" (UID: \"9ad4315c-9d47-485c-b656-4ba2e41803f6\") " pod="openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k" Nov 25 17:05:43 crc kubenswrapper[4802]: I1125 17:05:43.099645 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9ad4315c-9d47-485c-b656-4ba2e41803f6-util\") pod \"19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k\" (UID: \"9ad4315c-9d47-485c-b656-4ba2e41803f6\") " pod="openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k" Nov 25 17:05:43 crc kubenswrapper[4802]: I1125 17:05:43.099717 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7qst\" (UniqueName: \"kubernetes.io/projected/9ad4315c-9d47-485c-b656-4ba2e41803f6-kube-api-access-j7qst\") pod \"19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k\" (UID: \"9ad4315c-9d47-485c-b656-4ba2e41803f6\") " pod="openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k" Nov 25 17:05:43 crc kubenswrapper[4802]: I1125 17:05:43.099743 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9ad4315c-9d47-485c-b656-4ba2e41803f6-bundle\") pod \"19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k\" (UID: \"9ad4315c-9d47-485c-b656-4ba2e41803f6\") " pod="openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k" Nov 25 17:05:43 crc kubenswrapper[4802]: I1125 17:05:43.100382 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9ad4315c-9d47-485c-b656-4ba2e41803f6-util\") pod \"19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k\" (UID: \"9ad4315c-9d47-485c-b656-4ba2e41803f6\") " pod="openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k" Nov 25 17:05:43 crc kubenswrapper[4802]: I1125 17:05:43.100434 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9ad4315c-9d47-485c-b656-4ba2e41803f6-bundle\") pod \"19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k\" (UID: \"9ad4315c-9d47-485c-b656-4ba2e41803f6\") " pod="openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k" Nov 25 17:05:43 crc kubenswrapper[4802]: I1125 17:05:43.125858 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7qst\" (UniqueName: \"kubernetes.io/projected/9ad4315c-9d47-485c-b656-4ba2e41803f6-kube-api-access-j7qst\") pod \"19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k\" (UID: \"9ad4315c-9d47-485c-b656-4ba2e41803f6\") " pod="openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k" Nov 25 17:05:43 crc kubenswrapper[4802]: I1125 17:05:43.257503 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k" Nov 25 17:05:43 crc kubenswrapper[4802]: I1125 17:05:43.711971 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k"] Nov 25 17:05:44 crc kubenswrapper[4802]: I1125 17:05:44.472315 4802 generic.go:334] "Generic (PLEG): container finished" podID="9ad4315c-9d47-485c-b656-4ba2e41803f6" containerID="7419ec26e6c28776ebd607aec0202b20a283c13e1bd01d19afe056183948b9ed" exitCode=0 Nov 25 17:05:44 crc kubenswrapper[4802]: I1125 17:05:44.472366 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k" event={"ID":"9ad4315c-9d47-485c-b656-4ba2e41803f6","Type":"ContainerDied","Data":"7419ec26e6c28776ebd607aec0202b20a283c13e1bd01d19afe056183948b9ed"} Nov 25 17:05:44 crc kubenswrapper[4802]: I1125 17:05:44.472397 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k" event={"ID":"9ad4315c-9d47-485c-b656-4ba2e41803f6","Type":"ContainerStarted","Data":"94107411eae1296fdbe28fd9e695cc870b07d0af6981ff6509fbecc822021623"} Nov 25 17:05:45 crc kubenswrapper[4802]: I1125 17:05:45.481156 4802 generic.go:334] "Generic (PLEG): container finished" podID="9ad4315c-9d47-485c-b656-4ba2e41803f6" containerID="d0633eabec949525bb0479691f8cf609c30311dfe198ea69e0290a3d5bc01b01" exitCode=0 Nov 25 17:05:45 crc kubenswrapper[4802]: I1125 17:05:45.481203 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k" event={"ID":"9ad4315c-9d47-485c-b656-4ba2e41803f6","Type":"ContainerDied","Data":"d0633eabec949525bb0479691f8cf609c30311dfe198ea69e0290a3d5bc01b01"} Nov 25 17:05:46 crc kubenswrapper[4802]: I1125 17:05:46.490244 4802 generic.go:334] "Generic (PLEG): container finished" podID="9ad4315c-9d47-485c-b656-4ba2e41803f6" containerID="b3295dcbbcdc6261b443381c45747ae28e22f3f10675b219252bc3cc378a7c82" exitCode=0 Nov 25 17:05:46 crc kubenswrapper[4802]: I1125 17:05:46.490308 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k" event={"ID":"9ad4315c-9d47-485c-b656-4ba2e41803f6","Type":"ContainerDied","Data":"b3295dcbbcdc6261b443381c45747ae28e22f3f10675b219252bc3cc378a7c82"} Nov 25 17:05:47 crc kubenswrapper[4802]: I1125 17:05:47.741313 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k" Nov 25 17:05:47 crc kubenswrapper[4802]: I1125 17:05:47.867483 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7qst\" (UniqueName: \"kubernetes.io/projected/9ad4315c-9d47-485c-b656-4ba2e41803f6-kube-api-access-j7qst\") pod \"9ad4315c-9d47-485c-b656-4ba2e41803f6\" (UID: \"9ad4315c-9d47-485c-b656-4ba2e41803f6\") " Nov 25 17:05:47 crc kubenswrapper[4802]: I1125 17:05:47.867648 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9ad4315c-9d47-485c-b656-4ba2e41803f6-bundle\") pod \"9ad4315c-9d47-485c-b656-4ba2e41803f6\" (UID: \"9ad4315c-9d47-485c-b656-4ba2e41803f6\") " Nov 25 17:05:47 crc kubenswrapper[4802]: I1125 17:05:47.867698 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9ad4315c-9d47-485c-b656-4ba2e41803f6-util\") pod \"9ad4315c-9d47-485c-b656-4ba2e41803f6\" (UID: \"9ad4315c-9d47-485c-b656-4ba2e41803f6\") " Nov 25 17:05:47 crc kubenswrapper[4802]: I1125 17:05:47.868583 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ad4315c-9d47-485c-b656-4ba2e41803f6-bundle" (OuterVolumeSpecName: "bundle") pod "9ad4315c-9d47-485c-b656-4ba2e41803f6" (UID: "9ad4315c-9d47-485c-b656-4ba2e41803f6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:05:47 crc kubenswrapper[4802]: I1125 17:05:47.873284 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ad4315c-9d47-485c-b656-4ba2e41803f6-kube-api-access-j7qst" (OuterVolumeSpecName: "kube-api-access-j7qst") pod "9ad4315c-9d47-485c-b656-4ba2e41803f6" (UID: "9ad4315c-9d47-485c-b656-4ba2e41803f6"). InnerVolumeSpecName "kube-api-access-j7qst". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:05:47 crc kubenswrapper[4802]: I1125 17:05:47.881900 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ad4315c-9d47-485c-b656-4ba2e41803f6-util" (OuterVolumeSpecName: "util") pod "9ad4315c-9d47-485c-b656-4ba2e41803f6" (UID: "9ad4315c-9d47-485c-b656-4ba2e41803f6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:05:47 crc kubenswrapper[4802]: I1125 17:05:47.969041 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7qst\" (UniqueName: \"kubernetes.io/projected/9ad4315c-9d47-485c-b656-4ba2e41803f6-kube-api-access-j7qst\") on node \"crc\" DevicePath \"\"" Nov 25 17:05:47 crc kubenswrapper[4802]: I1125 17:05:47.969095 4802 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9ad4315c-9d47-485c-b656-4ba2e41803f6-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 17:05:47 crc kubenswrapper[4802]: I1125 17:05:47.969148 4802 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9ad4315c-9d47-485c-b656-4ba2e41803f6-util\") on node \"crc\" DevicePath \"\"" Nov 25 17:05:48 crc kubenswrapper[4802]: I1125 17:05:48.508898 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k" event={"ID":"9ad4315c-9d47-485c-b656-4ba2e41803f6","Type":"ContainerDied","Data":"94107411eae1296fdbe28fd9e695cc870b07d0af6981ff6509fbecc822021623"} Nov 25 17:05:48 crc kubenswrapper[4802]: I1125 17:05:48.509402 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94107411eae1296fdbe28fd9e695cc870b07d0af6981ff6509fbecc822021623" Nov 25 17:05:48 crc kubenswrapper[4802]: I1125 17:05:48.508978 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k" Nov 25 17:05:54 crc kubenswrapper[4802]: I1125 17:05:54.248022 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 17:05:54 crc kubenswrapper[4802]: I1125 17:05:54.248654 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 17:05:58 crc kubenswrapper[4802]: I1125 17:05:58.875785 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj"] Nov 25 17:05:58 crc kubenswrapper[4802]: E1125 17:05:58.876679 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ad4315c-9d47-485c-b656-4ba2e41803f6" containerName="util" Nov 25 17:05:58 crc kubenswrapper[4802]: I1125 17:05:58.876694 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ad4315c-9d47-485c-b656-4ba2e41803f6" containerName="util" Nov 25 17:05:58 crc kubenswrapper[4802]: E1125 17:05:58.876714 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ad4315c-9d47-485c-b656-4ba2e41803f6" containerName="pull" Nov 25 17:05:58 crc kubenswrapper[4802]: I1125 17:05:58.876722 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ad4315c-9d47-485c-b656-4ba2e41803f6" containerName="pull" Nov 25 17:05:58 crc kubenswrapper[4802]: E1125 17:05:58.876736 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ad4315c-9d47-485c-b656-4ba2e41803f6" containerName="extract" Nov 25 17:05:58 crc kubenswrapper[4802]: I1125 17:05:58.876742 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ad4315c-9d47-485c-b656-4ba2e41803f6" containerName="extract" Nov 25 17:05:58 crc kubenswrapper[4802]: I1125 17:05:58.876897 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ad4315c-9d47-485c-b656-4ba2e41803f6" containerName="extract" Nov 25 17:05:58 crc kubenswrapper[4802]: I1125 17:05:58.877493 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:05:58 crc kubenswrapper[4802]: I1125 17:05:58.879332 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Nov 25 17:05:58 crc kubenswrapper[4802]: I1125 17:05:58.879704 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-8pstl" Nov 25 17:05:58 crc kubenswrapper[4802]: I1125 17:05:58.902076 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj"] Nov 25 17:05:58 crc kubenswrapper[4802]: I1125 17:05:58.933063 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eeca972a-357e-4fa8-a8b5-1421b6601219-webhook-cert\") pod \"glance-operator-controller-manager-6bfbf5c856-xw6fj\" (UID: \"eeca972a-357e-4fa8-a8b5-1421b6601219\") " pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:05:58 crc kubenswrapper[4802]: I1125 17:05:58.933142 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eeca972a-357e-4fa8-a8b5-1421b6601219-apiservice-cert\") pod \"glance-operator-controller-manager-6bfbf5c856-xw6fj\" (UID: \"eeca972a-357e-4fa8-a8b5-1421b6601219\") " pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:05:58 crc kubenswrapper[4802]: I1125 17:05:58.933189 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbs7g\" (UniqueName: \"kubernetes.io/projected/eeca972a-357e-4fa8-a8b5-1421b6601219-kube-api-access-jbs7g\") pod \"glance-operator-controller-manager-6bfbf5c856-xw6fj\" (UID: \"eeca972a-357e-4fa8-a8b5-1421b6601219\") " pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:05:59 crc kubenswrapper[4802]: I1125 17:05:59.034339 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbs7g\" (UniqueName: \"kubernetes.io/projected/eeca972a-357e-4fa8-a8b5-1421b6601219-kube-api-access-jbs7g\") pod \"glance-operator-controller-manager-6bfbf5c856-xw6fj\" (UID: \"eeca972a-357e-4fa8-a8b5-1421b6601219\") " pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:05:59 crc kubenswrapper[4802]: I1125 17:05:59.034474 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eeca972a-357e-4fa8-a8b5-1421b6601219-webhook-cert\") pod \"glance-operator-controller-manager-6bfbf5c856-xw6fj\" (UID: \"eeca972a-357e-4fa8-a8b5-1421b6601219\") " pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:05:59 crc kubenswrapper[4802]: I1125 17:05:59.034504 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eeca972a-357e-4fa8-a8b5-1421b6601219-apiservice-cert\") pod \"glance-operator-controller-manager-6bfbf5c856-xw6fj\" (UID: \"eeca972a-357e-4fa8-a8b5-1421b6601219\") " pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:05:59 crc kubenswrapper[4802]: I1125 17:05:59.039802 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eeca972a-357e-4fa8-a8b5-1421b6601219-apiservice-cert\") pod \"glance-operator-controller-manager-6bfbf5c856-xw6fj\" (UID: \"eeca972a-357e-4fa8-a8b5-1421b6601219\") " pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:05:59 crc kubenswrapper[4802]: I1125 17:05:59.041730 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eeca972a-357e-4fa8-a8b5-1421b6601219-webhook-cert\") pod \"glance-operator-controller-manager-6bfbf5c856-xw6fj\" (UID: \"eeca972a-357e-4fa8-a8b5-1421b6601219\") " pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:05:59 crc kubenswrapper[4802]: I1125 17:05:59.057035 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbs7g\" (UniqueName: \"kubernetes.io/projected/eeca972a-357e-4fa8-a8b5-1421b6601219-kube-api-access-jbs7g\") pod \"glance-operator-controller-manager-6bfbf5c856-xw6fj\" (UID: \"eeca972a-357e-4fa8-a8b5-1421b6601219\") " pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:05:59 crc kubenswrapper[4802]: I1125 17:05:59.198584 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:05:59 crc kubenswrapper[4802]: I1125 17:05:59.638341 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj"] Nov 25 17:05:59 crc kubenswrapper[4802]: W1125 17:05:59.642950 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeeca972a_357e_4fa8_a8b5_1421b6601219.slice/crio-7eb83ca57c08f937754a2d44ae81d89739f95420e1e8414d5e28710232f0bc24 WatchSource:0}: Error finding container 7eb83ca57c08f937754a2d44ae81d89739f95420e1e8414d5e28710232f0bc24: Status 404 returned error can't find the container with id 7eb83ca57c08f937754a2d44ae81d89739f95420e1e8414d5e28710232f0bc24 Nov 25 17:06:00 crc kubenswrapper[4802]: I1125 17:06:00.597296 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" event={"ID":"eeca972a-357e-4fa8-a8b5-1421b6601219","Type":"ContainerStarted","Data":"7eb83ca57c08f937754a2d44ae81d89739f95420e1e8414d5e28710232f0bc24"} Nov 25 17:06:01 crc kubenswrapper[4802]: I1125 17:06:01.604942 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" event={"ID":"eeca972a-357e-4fa8-a8b5-1421b6601219","Type":"ContainerStarted","Data":"5fb925cd4ba59e71662af756f38cebc30afb1d1129d07ed4b4334801d2f89d4c"} Nov 25 17:06:01 crc kubenswrapper[4802]: I1125 17:06:01.606261 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:06:01 crc kubenswrapper[4802]: I1125 17:06:01.632606 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" podStartSLOduration=1.907952009 podStartE2EDuration="3.632574572s" podCreationTimestamp="2025-11-25 17:05:58 +0000 UTC" firstStartedPulling="2025-11-25 17:05:59.645389878 +0000 UTC m=+1142.789737064" lastFinishedPulling="2025-11-25 17:06:01.370012441 +0000 UTC m=+1144.514359627" observedRunningTime="2025-11-25 17:06:01.628571664 +0000 UTC m=+1144.772918850" watchObservedRunningTime="2025-11-25 17:06:01.632574572 +0000 UTC m=+1144.776921758" Nov 25 17:06:09 crc kubenswrapper[4802]: I1125 17:06:09.206456 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.023672 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-fxkc8"] Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.024975 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fxkc8" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.045002 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-fxkc8"] Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.055531 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znqzm\" (UniqueName: \"kubernetes.io/projected/ae47311c-06d0-4f16-a989-d0a60f7d7faa-kube-api-access-znqzm\") pod \"glance-db-create-fxkc8\" (UID: \"ae47311c-06d0-4f16-a989-d0a60f7d7faa\") " pod="glance-kuttl-tests/glance-db-create-fxkc8" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.055686 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae47311c-06d0-4f16-a989-d0a60f7d7faa-operator-scripts\") pod \"glance-db-create-fxkc8\" (UID: \"ae47311c-06d0-4f16-a989-d0a60f7d7faa\") " pod="glance-kuttl-tests/glance-db-create-fxkc8" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.109729 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.110616 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.112989 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.113414 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-vs6wx" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.113461 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.113641 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.127733 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.134669 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-ce03-account-create-update-ds8w8"] Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.138630 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-ce03-account-create-update-ds8w8" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.142423 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.157088 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-scripts\") pod \"openstackclient\" (UID: \"d9d79e72-cdd7-41be-886b-c72137db15bf\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.157182 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6rf5\" (UniqueName: \"kubernetes.io/projected/d9d79e72-cdd7-41be-886b-c72137db15bf-kube-api-access-t6rf5\") pod \"openstackclient\" (UID: \"d9d79e72-cdd7-41be-886b-c72137db15bf\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.157219 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44f61cc5-32b0-4f09-919a-c1d010c2ab28-operator-scripts\") pod \"glance-ce03-account-create-update-ds8w8\" (UID: \"44f61cc5-32b0-4f09-919a-c1d010c2ab28\") " pod="glance-kuttl-tests/glance-ce03-account-create-update-ds8w8" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.157271 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae47311c-06d0-4f16-a989-d0a60f7d7faa-operator-scripts\") pod \"glance-db-create-fxkc8\" (UID: \"ae47311c-06d0-4f16-a989-d0a60f7d7faa\") " pod="glance-kuttl-tests/glance-db-create-fxkc8" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.157296 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-config\") pod \"openstackclient\" (UID: \"d9d79e72-cdd7-41be-886b-c72137db15bf\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.157318 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd6wj\" (UniqueName: \"kubernetes.io/projected/44f61cc5-32b0-4f09-919a-c1d010c2ab28-kube-api-access-fd6wj\") pod \"glance-ce03-account-create-update-ds8w8\" (UID: \"44f61cc5-32b0-4f09-919a-c1d010c2ab28\") " pod="glance-kuttl-tests/glance-ce03-account-create-update-ds8w8" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.157349 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-config-secret\") pod \"openstackclient\" (UID: \"d9d79e72-cdd7-41be-886b-c72137db15bf\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.157395 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znqzm\" (UniqueName: \"kubernetes.io/projected/ae47311c-06d0-4f16-a989-d0a60f7d7faa-kube-api-access-znqzm\") pod \"glance-db-create-fxkc8\" (UID: \"ae47311c-06d0-4f16-a989-d0a60f7d7faa\") " pod="glance-kuttl-tests/glance-db-create-fxkc8" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.158521 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae47311c-06d0-4f16-a989-d0a60f7d7faa-operator-scripts\") pod \"glance-db-create-fxkc8\" (UID: \"ae47311c-06d0-4f16-a989-d0a60f7d7faa\") " pod="glance-kuttl-tests/glance-db-create-fxkc8" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.169617 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-ce03-account-create-update-ds8w8"] Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.178194 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znqzm\" (UniqueName: \"kubernetes.io/projected/ae47311c-06d0-4f16-a989-d0a60f7d7faa-kube-api-access-znqzm\") pod \"glance-db-create-fxkc8\" (UID: \"ae47311c-06d0-4f16-a989-d0a60f7d7faa\") " pod="glance-kuttl-tests/glance-db-create-fxkc8" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.258410 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6rf5\" (UniqueName: \"kubernetes.io/projected/d9d79e72-cdd7-41be-886b-c72137db15bf-kube-api-access-t6rf5\") pod \"openstackclient\" (UID: \"d9d79e72-cdd7-41be-886b-c72137db15bf\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.258463 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44f61cc5-32b0-4f09-919a-c1d010c2ab28-operator-scripts\") pod \"glance-ce03-account-create-update-ds8w8\" (UID: \"44f61cc5-32b0-4f09-919a-c1d010c2ab28\") " pod="glance-kuttl-tests/glance-ce03-account-create-update-ds8w8" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.258510 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-config\") pod \"openstackclient\" (UID: \"d9d79e72-cdd7-41be-886b-c72137db15bf\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.258529 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd6wj\" (UniqueName: \"kubernetes.io/projected/44f61cc5-32b0-4f09-919a-c1d010c2ab28-kube-api-access-fd6wj\") pod \"glance-ce03-account-create-update-ds8w8\" (UID: \"44f61cc5-32b0-4f09-919a-c1d010c2ab28\") " pod="glance-kuttl-tests/glance-ce03-account-create-update-ds8w8" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.258551 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-config-secret\") pod \"openstackclient\" (UID: \"d9d79e72-cdd7-41be-886b-c72137db15bf\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.258592 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-scripts\") pod \"openstackclient\" (UID: \"d9d79e72-cdd7-41be-886b-c72137db15bf\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.259538 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44f61cc5-32b0-4f09-919a-c1d010c2ab28-operator-scripts\") pod \"glance-ce03-account-create-update-ds8w8\" (UID: \"44f61cc5-32b0-4f09-919a-c1d010c2ab28\") " pod="glance-kuttl-tests/glance-ce03-account-create-update-ds8w8" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.260244 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-scripts\") pod \"openstackclient\" (UID: \"d9d79e72-cdd7-41be-886b-c72137db15bf\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.260308 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-config\") pod \"openstackclient\" (UID: \"d9d79e72-cdd7-41be-886b-c72137db15bf\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.263485 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-config-secret\") pod \"openstackclient\" (UID: \"d9d79e72-cdd7-41be-886b-c72137db15bf\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.276811 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6rf5\" (UniqueName: \"kubernetes.io/projected/d9d79e72-cdd7-41be-886b-c72137db15bf-kube-api-access-t6rf5\") pod \"openstackclient\" (UID: \"d9d79e72-cdd7-41be-886b-c72137db15bf\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.280819 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd6wj\" (UniqueName: \"kubernetes.io/projected/44f61cc5-32b0-4f09-919a-c1d010c2ab28-kube-api-access-fd6wj\") pod \"glance-ce03-account-create-update-ds8w8\" (UID: \"44f61cc5-32b0-4f09-919a-c1d010c2ab28\") " pod="glance-kuttl-tests/glance-ce03-account-create-update-ds8w8" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.350677 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fxkc8" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.430190 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.455438 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-ce03-account-create-update-ds8w8" Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.601765 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-fxkc8"] Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.711419 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-fxkc8" event={"ID":"ae47311c-06d0-4f16-a989-d0a60f7d7faa","Type":"ContainerStarted","Data":"a08e6592aac7078282a8c4ca99e93c9957854a8053cbbe98a310ce09b569d4cf"} Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.865646 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 17:06:14 crc kubenswrapper[4802]: W1125 17:06:14.872733 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9d79e72_cdd7_41be_886b_c72137db15bf.slice/crio-2d607d882fd73b4e4a8f4ae6ab876f25e02524a51094516986f2c9b75ccd4b15 WatchSource:0}: Error finding container 2d607d882fd73b4e4a8f4ae6ab876f25e02524a51094516986f2c9b75ccd4b15: Status 404 returned error can't find the container with id 2d607d882fd73b4e4a8f4ae6ab876f25e02524a51094516986f2c9b75ccd4b15 Nov 25 17:06:14 crc kubenswrapper[4802]: I1125 17:06:14.917547 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-ce03-account-create-update-ds8w8"] Nov 25 17:06:14 crc kubenswrapper[4802]: W1125 17:06:14.919613 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44f61cc5_32b0_4f09_919a_c1d010c2ab28.slice/crio-75a96401c331bab9839e0e39ad63889f0af333e9ca1765aa60419fbefd34d0df WatchSource:0}: Error finding container 75a96401c331bab9839e0e39ad63889f0af333e9ca1765aa60419fbefd34d0df: Status 404 returned error can't find the container with id 75a96401c331bab9839e0e39ad63889f0af333e9ca1765aa60419fbefd34d0df Nov 25 17:06:15 crc kubenswrapper[4802]: I1125 17:06:15.720731 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-ce03-account-create-update-ds8w8" event={"ID":"44f61cc5-32b0-4f09-919a-c1d010c2ab28","Type":"ContainerStarted","Data":"bcb0d353804d606af76bda79abe8eb0205565231c182c508116e7f9b967a3671"} Nov 25 17:06:15 crc kubenswrapper[4802]: I1125 17:06:15.720989 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-ce03-account-create-update-ds8w8" event={"ID":"44f61cc5-32b0-4f09-919a-c1d010c2ab28","Type":"ContainerStarted","Data":"75a96401c331bab9839e0e39ad63889f0af333e9ca1765aa60419fbefd34d0df"} Nov 25 17:06:15 crc kubenswrapper[4802]: I1125 17:06:15.724927 4802 generic.go:334] "Generic (PLEG): container finished" podID="ae47311c-06d0-4f16-a989-d0a60f7d7faa" containerID="d73f32305be41b0056e46f742b1c53aeb8f76f9cb88a65a5cd53e0539f2353a4" exitCode=0 Nov 25 17:06:15 crc kubenswrapper[4802]: I1125 17:06:15.725056 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-fxkc8" event={"ID":"ae47311c-06d0-4f16-a989-d0a60f7d7faa","Type":"ContainerDied","Data":"d73f32305be41b0056e46f742b1c53aeb8f76f9cb88a65a5cd53e0539f2353a4"} Nov 25 17:06:15 crc kubenswrapper[4802]: I1125 17:06:15.726643 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"d9d79e72-cdd7-41be-886b-c72137db15bf","Type":"ContainerStarted","Data":"2d607d882fd73b4e4a8f4ae6ab876f25e02524a51094516986f2c9b75ccd4b15"} Nov 25 17:06:15 crc kubenswrapper[4802]: I1125 17:06:15.736467 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-ce03-account-create-update-ds8w8" podStartSLOduration=1.736445722 podStartE2EDuration="1.736445722s" podCreationTimestamp="2025-11-25 17:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:06:15.734434478 +0000 UTC m=+1158.878781664" watchObservedRunningTime="2025-11-25 17:06:15.736445722 +0000 UTC m=+1158.880792908" Nov 25 17:06:16 crc kubenswrapper[4802]: I1125 17:06:16.740400 4802 generic.go:334] "Generic (PLEG): container finished" podID="44f61cc5-32b0-4f09-919a-c1d010c2ab28" containerID="bcb0d353804d606af76bda79abe8eb0205565231c182c508116e7f9b967a3671" exitCode=0 Nov 25 17:06:16 crc kubenswrapper[4802]: I1125 17:06:16.740584 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-ce03-account-create-update-ds8w8" event={"ID":"44f61cc5-32b0-4f09-919a-c1d010c2ab28","Type":"ContainerDied","Data":"bcb0d353804d606af76bda79abe8eb0205565231c182c508116e7f9b967a3671"} Nov 25 17:06:17 crc kubenswrapper[4802]: I1125 17:06:17.062735 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fxkc8" Nov 25 17:06:17 crc kubenswrapper[4802]: I1125 17:06:17.105528 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae47311c-06d0-4f16-a989-d0a60f7d7faa-operator-scripts\") pod \"ae47311c-06d0-4f16-a989-d0a60f7d7faa\" (UID: \"ae47311c-06d0-4f16-a989-d0a60f7d7faa\") " Nov 25 17:06:17 crc kubenswrapper[4802]: I1125 17:06:17.105633 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znqzm\" (UniqueName: \"kubernetes.io/projected/ae47311c-06d0-4f16-a989-d0a60f7d7faa-kube-api-access-znqzm\") pod \"ae47311c-06d0-4f16-a989-d0a60f7d7faa\" (UID: \"ae47311c-06d0-4f16-a989-d0a60f7d7faa\") " Nov 25 17:06:17 crc kubenswrapper[4802]: I1125 17:06:17.106740 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae47311c-06d0-4f16-a989-d0a60f7d7faa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ae47311c-06d0-4f16-a989-d0a60f7d7faa" (UID: "ae47311c-06d0-4f16-a989-d0a60f7d7faa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:06:17 crc kubenswrapper[4802]: I1125 17:06:17.111418 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae47311c-06d0-4f16-a989-d0a60f7d7faa-kube-api-access-znqzm" (OuterVolumeSpecName: "kube-api-access-znqzm") pod "ae47311c-06d0-4f16-a989-d0a60f7d7faa" (UID: "ae47311c-06d0-4f16-a989-d0a60f7d7faa"). InnerVolumeSpecName "kube-api-access-znqzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:06:17 crc kubenswrapper[4802]: I1125 17:06:17.206905 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znqzm\" (UniqueName: \"kubernetes.io/projected/ae47311c-06d0-4f16-a989-d0a60f7d7faa-kube-api-access-znqzm\") on node \"crc\" DevicePath \"\"" Nov 25 17:06:17 crc kubenswrapper[4802]: I1125 17:06:17.206939 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae47311c-06d0-4f16-a989-d0a60f7d7faa-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:06:17 crc kubenswrapper[4802]: I1125 17:06:17.763032 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fxkc8" Nov 25 17:06:17 crc kubenswrapper[4802]: I1125 17:06:17.763619 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-fxkc8" event={"ID":"ae47311c-06d0-4f16-a989-d0a60f7d7faa","Type":"ContainerDied","Data":"a08e6592aac7078282a8c4ca99e93c9957854a8053cbbe98a310ce09b569d4cf"} Nov 25 17:06:17 crc kubenswrapper[4802]: I1125 17:06:17.763643 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a08e6592aac7078282a8c4ca99e93c9957854a8053cbbe98a310ce09b569d4cf" Nov 25 17:06:18 crc kubenswrapper[4802]: I1125 17:06:18.149464 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-ce03-account-create-update-ds8w8" Nov 25 17:06:18 crc kubenswrapper[4802]: I1125 17:06:18.220090 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fd6wj\" (UniqueName: \"kubernetes.io/projected/44f61cc5-32b0-4f09-919a-c1d010c2ab28-kube-api-access-fd6wj\") pod \"44f61cc5-32b0-4f09-919a-c1d010c2ab28\" (UID: \"44f61cc5-32b0-4f09-919a-c1d010c2ab28\") " Nov 25 17:06:18 crc kubenswrapper[4802]: I1125 17:06:18.220512 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44f61cc5-32b0-4f09-919a-c1d010c2ab28-operator-scripts\") pod \"44f61cc5-32b0-4f09-919a-c1d010c2ab28\" (UID: \"44f61cc5-32b0-4f09-919a-c1d010c2ab28\") " Nov 25 17:06:18 crc kubenswrapper[4802]: I1125 17:06:18.222088 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44f61cc5-32b0-4f09-919a-c1d010c2ab28-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "44f61cc5-32b0-4f09-919a-c1d010c2ab28" (UID: "44f61cc5-32b0-4f09-919a-c1d010c2ab28"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:06:18 crc kubenswrapper[4802]: I1125 17:06:18.226780 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44f61cc5-32b0-4f09-919a-c1d010c2ab28-kube-api-access-fd6wj" (OuterVolumeSpecName: "kube-api-access-fd6wj") pod "44f61cc5-32b0-4f09-919a-c1d010c2ab28" (UID: "44f61cc5-32b0-4f09-919a-c1d010c2ab28"). InnerVolumeSpecName "kube-api-access-fd6wj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:06:18 crc kubenswrapper[4802]: I1125 17:06:18.323012 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fd6wj\" (UniqueName: \"kubernetes.io/projected/44f61cc5-32b0-4f09-919a-c1d010c2ab28-kube-api-access-fd6wj\") on node \"crc\" DevicePath \"\"" Nov 25 17:06:18 crc kubenswrapper[4802]: I1125 17:06:18.323043 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44f61cc5-32b0-4f09-919a-c1d010c2ab28-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:06:18 crc kubenswrapper[4802]: I1125 17:06:18.778151 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-ce03-account-create-update-ds8w8" event={"ID":"44f61cc5-32b0-4f09-919a-c1d010c2ab28","Type":"ContainerDied","Data":"75a96401c331bab9839e0e39ad63889f0af333e9ca1765aa60419fbefd34d0df"} Nov 25 17:06:18 crc kubenswrapper[4802]: I1125 17:06:18.778476 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75a96401c331bab9839e0e39ad63889f0af333e9ca1765aa60419fbefd34d0df" Nov 25 17:06:18 crc kubenswrapper[4802]: I1125 17:06:18.778537 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-ce03-account-create-update-ds8w8" Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.314693 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-lspxq"] Nov 25 17:06:19 crc kubenswrapper[4802]: E1125 17:06:19.314955 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44f61cc5-32b0-4f09-919a-c1d010c2ab28" containerName="mariadb-account-create-update" Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.314965 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="44f61cc5-32b0-4f09-919a-c1d010c2ab28" containerName="mariadb-account-create-update" Nov 25 17:06:19 crc kubenswrapper[4802]: E1125 17:06:19.314985 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae47311c-06d0-4f16-a989-d0a60f7d7faa" containerName="mariadb-database-create" Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.314991 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae47311c-06d0-4f16-a989-d0a60f7d7faa" containerName="mariadb-database-create" Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.315998 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae47311c-06d0-4f16-a989-d0a60f7d7faa" containerName="mariadb-database-create" Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.316055 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="44f61cc5-32b0-4f09-919a-c1d010c2ab28" containerName="mariadb-account-create-update" Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.316666 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-lspxq" Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.323754 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-wgzgl" Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.324324 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.330653 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-lspxq"] Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.336798 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ab3f03b-f8da-4082-8ece-aa2d9375accf-db-sync-config-data\") pod \"glance-db-sync-lspxq\" (UID: \"5ab3f03b-f8da-4082-8ece-aa2d9375accf\") " pod="glance-kuttl-tests/glance-db-sync-lspxq" Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.336874 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ab3f03b-f8da-4082-8ece-aa2d9375accf-config-data\") pod \"glance-db-sync-lspxq\" (UID: \"5ab3f03b-f8da-4082-8ece-aa2d9375accf\") " pod="glance-kuttl-tests/glance-db-sync-lspxq" Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.336898 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt97h\" (UniqueName: \"kubernetes.io/projected/5ab3f03b-f8da-4082-8ece-aa2d9375accf-kube-api-access-vt97h\") pod \"glance-db-sync-lspxq\" (UID: \"5ab3f03b-f8da-4082-8ece-aa2d9375accf\") " pod="glance-kuttl-tests/glance-db-sync-lspxq" Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.437980 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ab3f03b-f8da-4082-8ece-aa2d9375accf-config-data\") pod \"glance-db-sync-lspxq\" (UID: \"5ab3f03b-f8da-4082-8ece-aa2d9375accf\") " pod="glance-kuttl-tests/glance-db-sync-lspxq" Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.438030 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt97h\" (UniqueName: \"kubernetes.io/projected/5ab3f03b-f8da-4082-8ece-aa2d9375accf-kube-api-access-vt97h\") pod \"glance-db-sync-lspxq\" (UID: \"5ab3f03b-f8da-4082-8ece-aa2d9375accf\") " pod="glance-kuttl-tests/glance-db-sync-lspxq" Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.438096 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ab3f03b-f8da-4082-8ece-aa2d9375accf-db-sync-config-data\") pod \"glance-db-sync-lspxq\" (UID: \"5ab3f03b-f8da-4082-8ece-aa2d9375accf\") " pod="glance-kuttl-tests/glance-db-sync-lspxq" Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.442708 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ab3f03b-f8da-4082-8ece-aa2d9375accf-db-sync-config-data\") pod \"glance-db-sync-lspxq\" (UID: \"5ab3f03b-f8da-4082-8ece-aa2d9375accf\") " pod="glance-kuttl-tests/glance-db-sync-lspxq" Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.453235 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ab3f03b-f8da-4082-8ece-aa2d9375accf-config-data\") pod \"glance-db-sync-lspxq\" (UID: \"5ab3f03b-f8da-4082-8ece-aa2d9375accf\") " pod="glance-kuttl-tests/glance-db-sync-lspxq" Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.462983 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt97h\" (UniqueName: \"kubernetes.io/projected/5ab3f03b-f8da-4082-8ece-aa2d9375accf-kube-api-access-vt97h\") pod \"glance-db-sync-lspxq\" (UID: \"5ab3f03b-f8da-4082-8ece-aa2d9375accf\") " pod="glance-kuttl-tests/glance-db-sync-lspxq" Nov 25 17:06:19 crc kubenswrapper[4802]: I1125 17:06:19.642048 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-lspxq" Nov 25 17:06:24 crc kubenswrapper[4802]: I1125 17:06:24.249091 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 17:06:24 crc kubenswrapper[4802]: I1125 17:06:24.249883 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 17:06:24 crc kubenswrapper[4802]: I1125 17:06:24.249959 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 17:06:24 crc kubenswrapper[4802]: I1125 17:06:24.250984 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"18e081a4af641b516662b68a934da8c320eec5e2196f1740999f260df1353261"} pod="openshift-machine-config-operator/machine-config-daemon-h29wc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 17:06:24 crc kubenswrapper[4802]: I1125 17:06:24.251077 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" containerID="cri-o://18e081a4af641b516662b68a934da8c320eec5e2196f1740999f260df1353261" gracePeriod=600 Nov 25 17:06:24 crc kubenswrapper[4802]: I1125 17:06:24.292528 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-lspxq"] Nov 25 17:06:24 crc kubenswrapper[4802]: I1125 17:06:24.857567 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-lspxq" event={"ID":"5ab3f03b-f8da-4082-8ece-aa2d9375accf","Type":"ContainerStarted","Data":"81625b49f16a465e3ea59b12cef4489200cb0d3f88bb65d101465780f98951e8"} Nov 25 17:06:24 crc kubenswrapper[4802]: I1125 17:06:24.859297 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"d9d79e72-cdd7-41be-886b-c72137db15bf","Type":"ContainerStarted","Data":"5a791aa521fc8c3c6a0d04b3916489ef736afb9b3a9fcf3c35b6a9c60c7e0e42"} Nov 25 17:06:24 crc kubenswrapper[4802]: I1125 17:06:24.862257 4802 generic.go:334] "Generic (PLEG): container finished" podID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerID="18e081a4af641b516662b68a934da8c320eec5e2196f1740999f260df1353261" exitCode=0 Nov 25 17:06:24 crc kubenswrapper[4802]: I1125 17:06:24.862304 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerDied","Data":"18e081a4af641b516662b68a934da8c320eec5e2196f1740999f260df1353261"} Nov 25 17:06:24 crc kubenswrapper[4802]: I1125 17:06:24.862330 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerStarted","Data":"ee1883bc854fb1518fe7b71c8f9bba592e9827245176f86eb7e6da8b81dacf06"} Nov 25 17:06:24 crc kubenswrapper[4802]: I1125 17:06:24.862346 4802 scope.go:117] "RemoveContainer" containerID="b903688e55473d1e5bdabb3f4b23a6b1ac78c87eaabad518c885de28866fe47e" Nov 25 17:06:24 crc kubenswrapper[4802]: I1125 17:06:24.880648 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=1.541285483 podStartE2EDuration="10.880628168s" podCreationTimestamp="2025-11-25 17:06:14 +0000 UTC" firstStartedPulling="2025-11-25 17:06:14.874554927 +0000 UTC m=+1158.018902113" lastFinishedPulling="2025-11-25 17:06:24.213897582 +0000 UTC m=+1167.358244798" observedRunningTime="2025-11-25 17:06:24.872889539 +0000 UTC m=+1168.017236725" watchObservedRunningTime="2025-11-25 17:06:24.880628168 +0000 UTC m=+1168.024975354" Nov 25 17:06:37 crc kubenswrapper[4802]: I1125 17:06:37.970626 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-lspxq" event={"ID":"5ab3f03b-f8da-4082-8ece-aa2d9375accf","Type":"ContainerStarted","Data":"faf65e25c9f071d1e0ebfb2348654e108cf3fbc584391de072ac29e6b7831a80"} Nov 25 17:06:37 crc kubenswrapper[4802]: I1125 17:06:37.996256 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-lspxq" podStartSLOduration=6.668759872 podStartE2EDuration="18.996225811s" podCreationTimestamp="2025-11-25 17:06:19 +0000 UTC" firstStartedPulling="2025-11-25 17:06:24.308013375 +0000 UTC m=+1167.452360571" lastFinishedPulling="2025-11-25 17:06:36.635479324 +0000 UTC m=+1179.779826510" observedRunningTime="2025-11-25 17:06:37.989884339 +0000 UTC m=+1181.134231535" watchObservedRunningTime="2025-11-25 17:06:37.996225811 +0000 UTC m=+1181.140572997" Nov 25 17:06:49 crc kubenswrapper[4802]: I1125 17:06:49.049922 4802 generic.go:334] "Generic (PLEG): container finished" podID="5ab3f03b-f8da-4082-8ece-aa2d9375accf" containerID="faf65e25c9f071d1e0ebfb2348654e108cf3fbc584391de072ac29e6b7831a80" exitCode=0 Nov 25 17:06:49 crc kubenswrapper[4802]: I1125 17:06:49.050050 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-lspxq" event={"ID":"5ab3f03b-f8da-4082-8ece-aa2d9375accf","Type":"ContainerDied","Data":"faf65e25c9f071d1e0ebfb2348654e108cf3fbc584391de072ac29e6b7831a80"} Nov 25 17:06:50 crc kubenswrapper[4802]: I1125 17:06:50.308701 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-lspxq" Nov 25 17:06:50 crc kubenswrapper[4802]: I1125 17:06:50.460091 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt97h\" (UniqueName: \"kubernetes.io/projected/5ab3f03b-f8da-4082-8ece-aa2d9375accf-kube-api-access-vt97h\") pod \"5ab3f03b-f8da-4082-8ece-aa2d9375accf\" (UID: \"5ab3f03b-f8da-4082-8ece-aa2d9375accf\") " Nov 25 17:06:50 crc kubenswrapper[4802]: I1125 17:06:50.460256 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ab3f03b-f8da-4082-8ece-aa2d9375accf-db-sync-config-data\") pod \"5ab3f03b-f8da-4082-8ece-aa2d9375accf\" (UID: \"5ab3f03b-f8da-4082-8ece-aa2d9375accf\") " Nov 25 17:06:50 crc kubenswrapper[4802]: I1125 17:06:50.460349 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ab3f03b-f8da-4082-8ece-aa2d9375accf-config-data\") pod \"5ab3f03b-f8da-4082-8ece-aa2d9375accf\" (UID: \"5ab3f03b-f8da-4082-8ece-aa2d9375accf\") " Nov 25 17:06:50 crc kubenswrapper[4802]: I1125 17:06:50.467959 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ab3f03b-f8da-4082-8ece-aa2d9375accf-kube-api-access-vt97h" (OuterVolumeSpecName: "kube-api-access-vt97h") pod "5ab3f03b-f8da-4082-8ece-aa2d9375accf" (UID: "5ab3f03b-f8da-4082-8ece-aa2d9375accf"). InnerVolumeSpecName "kube-api-access-vt97h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:06:50 crc kubenswrapper[4802]: I1125 17:06:50.467997 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ab3f03b-f8da-4082-8ece-aa2d9375accf-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5ab3f03b-f8da-4082-8ece-aa2d9375accf" (UID: "5ab3f03b-f8da-4082-8ece-aa2d9375accf"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:06:50 crc kubenswrapper[4802]: I1125 17:06:50.510721 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ab3f03b-f8da-4082-8ece-aa2d9375accf-config-data" (OuterVolumeSpecName: "config-data") pod "5ab3f03b-f8da-4082-8ece-aa2d9375accf" (UID: "5ab3f03b-f8da-4082-8ece-aa2d9375accf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:06:50 crc kubenswrapper[4802]: I1125 17:06:50.563028 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt97h\" (UniqueName: \"kubernetes.io/projected/5ab3f03b-f8da-4082-8ece-aa2d9375accf-kube-api-access-vt97h\") on node \"crc\" DevicePath \"\"" Nov 25 17:06:50 crc kubenswrapper[4802]: I1125 17:06:50.563575 4802 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ab3f03b-f8da-4082-8ece-aa2d9375accf-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 17:06:50 crc kubenswrapper[4802]: I1125 17:06:50.563650 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ab3f03b-f8da-4082-8ece-aa2d9375accf-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 17:06:51 crc kubenswrapper[4802]: I1125 17:06:51.068677 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-lspxq" event={"ID":"5ab3f03b-f8da-4082-8ece-aa2d9375accf","Type":"ContainerDied","Data":"81625b49f16a465e3ea59b12cef4489200cb0d3f88bb65d101465780f98951e8"} Nov 25 17:06:51 crc kubenswrapper[4802]: I1125 17:06:51.068743 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81625b49f16a465e3ea59b12cef4489200cb0d3f88bb65d101465780f98951e8" Nov 25 17:06:51 crc kubenswrapper[4802]: I1125 17:06:51.068772 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-lspxq" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.420041 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 17:06:52 crc kubenswrapper[4802]: E1125 17:06:52.420497 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ab3f03b-f8da-4082-8ece-aa2d9375accf" containerName="glance-db-sync" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.420519 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ab3f03b-f8da-4082-8ece-aa2d9375accf" containerName="glance-db-sync" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.420769 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ab3f03b-f8da-4082-8ece-aa2d9375accf" containerName="glance-db-sync" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.422028 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.424321 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.424362 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.431225 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-wgzgl" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.435234 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.491506 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.493082 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.504205 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.592638 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-dev\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.592689 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-dev\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.592708 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-sys\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.592728 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e244ce8-d122-4a53-bc3e-e42ce7231332-config-data\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.592934 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.592992 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-lib-modules\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.593044 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2f664f6-2086-40bb-9cfe-0544dfabd571-httpd-run\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.593135 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-etc-nvme\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.593215 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.593248 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2f664f6-2086-40bb-9cfe-0544dfabd571-logs\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.593266 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2f664f6-2086-40bb-9cfe-0544dfabd571-scripts\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.593325 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-etc-nvme\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.593360 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-run\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.593417 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.593485 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e244ce8-d122-4a53-bc3e-e42ce7231332-logs\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.593545 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-sys\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.593595 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7e244ce8-d122-4a53-bc3e-e42ce7231332-httpd-run\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.593642 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwvhm\" (UniqueName: \"kubernetes.io/projected/7e244ce8-d122-4a53-bc3e-e42ce7231332-kube-api-access-mwvhm\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.593665 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.593887 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e244ce8-d122-4a53-bc3e-e42ce7231332-scripts\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.593954 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-run\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.594383 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2f664f6-2086-40bb-9cfe-0544dfabd571-config-data\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.594427 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.594545 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.594644 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-lib-modules\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.594719 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vp99\" (UniqueName: \"kubernetes.io/projected/f2f664f6-2086-40bb-9cfe-0544dfabd571-kube-api-access-2vp99\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.594775 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.594814 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.703840 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.703925 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.703976 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-dev\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704001 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-sys\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704026 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e244ce8-d122-4a53-bc3e-e42ce7231332-config-data\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704048 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-dev\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704074 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704101 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-lib-modules\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704155 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2f664f6-2086-40bb-9cfe-0544dfabd571-httpd-run\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704190 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-etc-nvme\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704250 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704294 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2f664f6-2086-40bb-9cfe-0544dfabd571-logs\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704329 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2f664f6-2086-40bb-9cfe-0544dfabd571-scripts\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704382 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-etc-nvme\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704415 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-run\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704468 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704523 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e244ce8-d122-4a53-bc3e-e42ce7231332-logs\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704564 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-sys\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704598 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7e244ce8-d122-4a53-bc3e-e42ce7231332-httpd-run\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704630 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwvhm\" (UniqueName: \"kubernetes.io/projected/7e244ce8-d122-4a53-bc3e-e42ce7231332-kube-api-access-mwvhm\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704653 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704678 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e244ce8-d122-4a53-bc3e-e42ce7231332-scripts\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704720 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-run\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704758 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2f664f6-2086-40bb-9cfe-0544dfabd571-config-data\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704787 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704815 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704841 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-lib-modules\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.704866 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vp99\" (UniqueName: \"kubernetes.io/projected/f2f664f6-2086-40bb-9cfe-0544dfabd571-kube-api-access-2vp99\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.705627 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.709607 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-etc-nvme\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.709704 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.709754 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-sys\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.710206 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7e244ce8-d122-4a53-bc3e-e42ce7231332-httpd-run\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.710259 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-run\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.710336 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-lib-modules\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.710466 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-dev\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.710511 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.710568 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.710637 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.710681 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-etc-nvme\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.710685 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-dev\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.710737 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-run\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.710745 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-lib-modules\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.710784 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-sys\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.710847 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.711257 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2f664f6-2086-40bb-9cfe-0544dfabd571-httpd-run\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.711279 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2f664f6-2086-40bb-9cfe-0544dfabd571-logs\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.711332 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.711366 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.711763 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e244ce8-d122-4a53-bc3e-e42ce7231332-logs\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.719865 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e244ce8-d122-4a53-bc3e-e42ce7231332-scripts\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.735637 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2f664f6-2086-40bb-9cfe-0544dfabd571-scripts\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.746365 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwvhm\" (UniqueName: \"kubernetes.io/projected/7e244ce8-d122-4a53-bc3e-e42ce7231332-kube-api-access-mwvhm\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.749351 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2f664f6-2086-40bb-9cfe-0544dfabd571-config-data\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.749401 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e244ce8-d122-4a53-bc3e-e42ce7231332-config-data\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.751438 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vp99\" (UniqueName: \"kubernetes.io/projected/f2f664f6-2086-40bb-9cfe-0544dfabd571-kube-api-access-2vp99\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.755049 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.755751 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.761680 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.767457 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:52 crc kubenswrapper[4802]: I1125 17:06:52.807005 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:06:53 crc kubenswrapper[4802]: I1125 17:06:53.039603 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:06:54 crc kubenswrapper[4802]: I1125 17:06:54.806963 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 17:06:55 crc kubenswrapper[4802]: I1125 17:06:55.083199 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 17:06:55 crc kubenswrapper[4802]: I1125 17:06:55.115846 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7e244ce8-d122-4a53-bc3e-e42ce7231332","Type":"ContainerStarted","Data":"9948232ee67b000c1837da9dd885c8b1a6db451e1932ea182fefe3c19cc29d84"} Nov 25 17:06:55 crc kubenswrapper[4802]: I1125 17:06:55.117405 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"f2f664f6-2086-40bb-9cfe-0544dfabd571","Type":"ContainerStarted","Data":"2eeebce12b45fe68cf89f29b7bb76923c043a720b6e2c914589f09b3f073849c"} Nov 25 17:06:56 crc kubenswrapper[4802]: I1125 17:06:56.136677 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"f2f664f6-2086-40bb-9cfe-0544dfabd571","Type":"ContainerStarted","Data":"239b5f614d0559a9c8d73d157796461e709207d744d4d00f57888b782d3f328c"} Nov 25 17:06:56 crc kubenswrapper[4802]: I1125 17:06:56.139743 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7e244ce8-d122-4a53-bc3e-e42ce7231332","Type":"ContainerStarted","Data":"91c7e79258e4335d0e26ebf1f1fdcd545be2c5887d837c138681d45f5e5dfe5b"} Nov 25 17:06:56 crc kubenswrapper[4802]: I1125 17:06:56.139791 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7e244ce8-d122-4a53-bc3e-e42ce7231332","Type":"ContainerStarted","Data":"a0d534cfcfe30ecbeea0b44c88b4133fa6b764a6c5cfa131608459ddd7f9e32c"} Nov 25 17:06:56 crc kubenswrapper[4802]: I1125 17:06:56.161354 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=5.161333334 podStartE2EDuration="5.161333334s" podCreationTimestamp="2025-11-25 17:06:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:06:56.160908822 +0000 UTC m=+1199.305256008" watchObservedRunningTime="2025-11-25 17:06:56.161333334 +0000 UTC m=+1199.305680530" Nov 25 17:06:57 crc kubenswrapper[4802]: I1125 17:06:57.150274 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"f2f664f6-2086-40bb-9cfe-0544dfabd571","Type":"ContainerStarted","Data":"3208ac757d732c9a0b5234760fd13ad44ecd4525c8b4cd547cf8bfcf6f04cd61"} Nov 25 17:06:57 crc kubenswrapper[4802]: I1125 17:06:57.179104 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=5.179074618 podStartE2EDuration="5.179074618s" podCreationTimestamp="2025-11-25 17:06:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:06:57.172560192 +0000 UTC m=+1200.316907458" watchObservedRunningTime="2025-11-25 17:06:57.179074618 +0000 UTC m=+1200.323421844" Nov 25 17:07:02 crc kubenswrapper[4802]: I1125 17:07:02.807557 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:02 crc kubenswrapper[4802]: I1125 17:07:02.808166 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:02 crc kubenswrapper[4802]: I1125 17:07:02.836575 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:02 crc kubenswrapper[4802]: I1125 17:07:02.858561 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:03 crc kubenswrapper[4802]: I1125 17:07:03.040383 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:07:03 crc kubenswrapper[4802]: I1125 17:07:03.040509 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:07:03 crc kubenswrapper[4802]: I1125 17:07:03.089772 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:07:03 crc kubenswrapper[4802]: I1125 17:07:03.096533 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:07:03 crc kubenswrapper[4802]: I1125 17:07:03.206342 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:07:03 crc kubenswrapper[4802]: I1125 17:07:03.206778 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:03 crc kubenswrapper[4802]: I1125 17:07:03.206791 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:03 crc kubenswrapper[4802]: I1125 17:07:03.206799 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:07:05 crc kubenswrapper[4802]: I1125 17:07:05.142317 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:05 crc kubenswrapper[4802]: I1125 17:07:05.191688 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:07:05 crc kubenswrapper[4802]: I1125 17:07:05.192965 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:07:05 crc kubenswrapper[4802]: I1125 17:07:05.225949 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 17:07:05 crc kubenswrapper[4802]: I1125 17:07:05.235204 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:05 crc kubenswrapper[4802]: I1125 17:07:05.268797 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 17:07:07 crc kubenswrapper[4802]: I1125 17:07:07.243085 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="7e244ce8-d122-4a53-bc3e-e42ce7231332" containerName="glance-log" containerID="cri-o://a0d534cfcfe30ecbeea0b44c88b4133fa6b764a6c5cfa131608459ddd7f9e32c" gracePeriod=30 Nov 25 17:07:07 crc kubenswrapper[4802]: I1125 17:07:07.243240 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="7e244ce8-d122-4a53-bc3e-e42ce7231332" containerName="glance-httpd" containerID="cri-o://91c7e79258e4335d0e26ebf1f1fdcd545be2c5887d837c138681d45f5e5dfe5b" gracePeriod=30 Nov 25 17:07:08 crc kubenswrapper[4802]: I1125 17:07:08.257812 4802 generic.go:334] "Generic (PLEG): container finished" podID="7e244ce8-d122-4a53-bc3e-e42ce7231332" containerID="a0d534cfcfe30ecbeea0b44c88b4133fa6b764a6c5cfa131608459ddd7f9e32c" exitCode=143 Nov 25 17:07:08 crc kubenswrapper[4802]: I1125 17:07:08.258247 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7e244ce8-d122-4a53-bc3e-e42ce7231332","Type":"ContainerDied","Data":"a0d534cfcfe30ecbeea0b44c88b4133fa6b764a6c5cfa131608459ddd7f9e32c"} Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.784064 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.806855 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e244ce8-d122-4a53-bc3e-e42ce7231332-config-data\") pod \"7e244ce8-d122-4a53-bc3e-e42ce7231332\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.807210 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"7e244ce8-d122-4a53-bc3e-e42ce7231332\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.807389 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-run\") pod \"7e244ce8-d122-4a53-bc3e-e42ce7231332\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.807546 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-sys\") pod \"7e244ce8-d122-4a53-bc3e-e42ce7231332\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.807738 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e244ce8-d122-4a53-bc3e-e42ce7231332-logs\") pod \"7e244ce8-d122-4a53-bc3e-e42ce7231332\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.807581 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-run" (OuterVolumeSpecName: "run") pod "7e244ce8-d122-4a53-bc3e-e42ce7231332" (UID: "7e244ce8-d122-4a53-bc3e-e42ce7231332"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.807632 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-sys" (OuterVolumeSpecName: "sys") pod "7e244ce8-d122-4a53-bc3e-e42ce7231332" (UID: "7e244ce8-d122-4a53-bc3e-e42ce7231332"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.807955 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-etc-nvme\") pod \"7e244ce8-d122-4a53-bc3e-e42ce7231332\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.808111 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7e244ce8-d122-4a53-bc3e-e42ce7231332-httpd-run\") pod \"7e244ce8-d122-4a53-bc3e-e42ce7231332\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.808115 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e244ce8-d122-4a53-bc3e-e42ce7231332-logs" (OuterVolumeSpecName: "logs") pod "7e244ce8-d122-4a53-bc3e-e42ce7231332" (UID: "7e244ce8-d122-4a53-bc3e-e42ce7231332"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.808193 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"7e244ce8-d122-4a53-bc3e-e42ce7231332\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.808232 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-lib-modules\") pod \"7e244ce8-d122-4a53-bc3e-e42ce7231332\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.808292 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-dev\") pod \"7e244ce8-d122-4a53-bc3e-e42ce7231332\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.808327 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "7e244ce8-d122-4a53-bc3e-e42ce7231332" (UID: "7e244ce8-d122-4a53-bc3e-e42ce7231332"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.808377 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e244ce8-d122-4a53-bc3e-e42ce7231332-scripts\") pod \"7e244ce8-d122-4a53-bc3e-e42ce7231332\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.808413 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-dev" (OuterVolumeSpecName: "dev") pod "7e244ce8-d122-4a53-bc3e-e42ce7231332" (UID: "7e244ce8-d122-4a53-bc3e-e42ce7231332"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.808452 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-etc-iscsi\") pod \"7e244ce8-d122-4a53-bc3e-e42ce7231332\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.808530 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwvhm\" (UniqueName: \"kubernetes.io/projected/7e244ce8-d122-4a53-bc3e-e42ce7231332-kube-api-access-mwvhm\") pod \"7e244ce8-d122-4a53-bc3e-e42ce7231332\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.808564 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-var-locks-brick\") pod \"7e244ce8-d122-4a53-bc3e-e42ce7231332\" (UID: \"7e244ce8-d122-4a53-bc3e-e42ce7231332\") " Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.808843 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "7e244ce8-d122-4a53-bc3e-e42ce7231332" (UID: "7e244ce8-d122-4a53-bc3e-e42ce7231332"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.808869 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e244ce8-d122-4a53-bc3e-e42ce7231332-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7e244ce8-d122-4a53-bc3e-e42ce7231332" (UID: "7e244ce8-d122-4a53-bc3e-e42ce7231332"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.808908 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "7e244ce8-d122-4a53-bc3e-e42ce7231332" (UID: "7e244ce8-d122-4a53-bc3e-e42ce7231332"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.809198 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "7e244ce8-d122-4a53-bc3e-e42ce7231332" (UID: "7e244ce8-d122-4a53-bc3e-e42ce7231332"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.809426 4802 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-dev\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.809462 4802 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.809479 4802 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.809492 4802 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.809504 4802 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-sys\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.809517 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e244ce8-d122-4a53-bc3e-e42ce7231332-logs\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.809529 4802 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7e244ce8-d122-4a53-bc3e-e42ce7231332-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.809542 4802 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.821874 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "7e244ce8-d122-4a53-bc3e-e42ce7231332" (UID: "7e244ce8-d122-4a53-bc3e-e42ce7231332"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.823020 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e244ce8-d122-4a53-bc3e-e42ce7231332-scripts" (OuterVolumeSpecName: "scripts") pod "7e244ce8-d122-4a53-bc3e-e42ce7231332" (UID: "7e244ce8-d122-4a53-bc3e-e42ce7231332"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.823985 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "7e244ce8-d122-4a53-bc3e-e42ce7231332" (UID: "7e244ce8-d122-4a53-bc3e-e42ce7231332"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.827898 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e244ce8-d122-4a53-bc3e-e42ce7231332-kube-api-access-mwvhm" (OuterVolumeSpecName: "kube-api-access-mwvhm") pod "7e244ce8-d122-4a53-bc3e-e42ce7231332" (UID: "7e244ce8-d122-4a53-bc3e-e42ce7231332"). InnerVolumeSpecName "kube-api-access-mwvhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.860290 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e244ce8-d122-4a53-bc3e-e42ce7231332-config-data" (OuterVolumeSpecName: "config-data") pod "7e244ce8-d122-4a53-bc3e-e42ce7231332" (UID: "7e244ce8-d122-4a53-bc3e-e42ce7231332"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.912045 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e244ce8-d122-4a53-bc3e-e42ce7231332-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.912107 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwvhm\" (UniqueName: \"kubernetes.io/projected/7e244ce8-d122-4a53-bc3e-e42ce7231332-kube-api-access-mwvhm\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.912165 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e244ce8-d122-4a53-bc3e-e42ce7231332-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.912350 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.912369 4802 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7e244ce8-d122-4a53-bc3e-e42ce7231332-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.912393 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.927547 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 25 17:07:10 crc kubenswrapper[4802]: I1125 17:07:10.937705 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.014453 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.014502 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.287623 4802 generic.go:334] "Generic (PLEG): container finished" podID="7e244ce8-d122-4a53-bc3e-e42ce7231332" containerID="91c7e79258e4335d0e26ebf1f1fdcd545be2c5887d837c138681d45f5e5dfe5b" exitCode=0 Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.287696 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7e244ce8-d122-4a53-bc3e-e42ce7231332","Type":"ContainerDied","Data":"91c7e79258e4335d0e26ebf1f1fdcd545be2c5887d837c138681d45f5e5dfe5b"} Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.287760 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7e244ce8-d122-4a53-bc3e-e42ce7231332","Type":"ContainerDied","Data":"9948232ee67b000c1837da9dd885c8b1a6db451e1932ea182fefe3c19cc29d84"} Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.287762 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.287783 4802 scope.go:117] "RemoveContainer" containerID="91c7e79258e4335d0e26ebf1f1fdcd545be2c5887d837c138681d45f5e5dfe5b" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.336233 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.338955 4802 scope.go:117] "RemoveContainer" containerID="a0d534cfcfe30ecbeea0b44c88b4133fa6b764a6c5cfa131608459ddd7f9e32c" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.344869 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.357439 4802 scope.go:117] "RemoveContainer" containerID="91c7e79258e4335d0e26ebf1f1fdcd545be2c5887d837c138681d45f5e5dfe5b" Nov 25 17:07:11 crc kubenswrapper[4802]: E1125 17:07:11.364214 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91c7e79258e4335d0e26ebf1f1fdcd545be2c5887d837c138681d45f5e5dfe5b\": container with ID starting with 91c7e79258e4335d0e26ebf1f1fdcd545be2c5887d837c138681d45f5e5dfe5b not found: ID does not exist" containerID="91c7e79258e4335d0e26ebf1f1fdcd545be2c5887d837c138681d45f5e5dfe5b" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.364257 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91c7e79258e4335d0e26ebf1f1fdcd545be2c5887d837c138681d45f5e5dfe5b"} err="failed to get container status \"91c7e79258e4335d0e26ebf1f1fdcd545be2c5887d837c138681d45f5e5dfe5b\": rpc error: code = NotFound desc = could not find container \"91c7e79258e4335d0e26ebf1f1fdcd545be2c5887d837c138681d45f5e5dfe5b\": container with ID starting with 91c7e79258e4335d0e26ebf1f1fdcd545be2c5887d837c138681d45f5e5dfe5b not found: ID does not exist" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.364285 4802 scope.go:117] "RemoveContainer" containerID="a0d534cfcfe30ecbeea0b44c88b4133fa6b764a6c5cfa131608459ddd7f9e32c" Nov 25 17:07:11 crc kubenswrapper[4802]: E1125 17:07:11.364723 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0d534cfcfe30ecbeea0b44c88b4133fa6b764a6c5cfa131608459ddd7f9e32c\": container with ID starting with a0d534cfcfe30ecbeea0b44c88b4133fa6b764a6c5cfa131608459ddd7f9e32c not found: ID does not exist" containerID="a0d534cfcfe30ecbeea0b44c88b4133fa6b764a6c5cfa131608459ddd7f9e32c" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.364744 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0d534cfcfe30ecbeea0b44c88b4133fa6b764a6c5cfa131608459ddd7f9e32c"} err="failed to get container status \"a0d534cfcfe30ecbeea0b44c88b4133fa6b764a6c5cfa131608459ddd7f9e32c\": rpc error: code = NotFound desc = could not find container \"a0d534cfcfe30ecbeea0b44c88b4133fa6b764a6c5cfa131608459ddd7f9e32c\": container with ID starting with a0d534cfcfe30ecbeea0b44c88b4133fa6b764a6c5cfa131608459ddd7f9e32c not found: ID does not exist" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.368318 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 17:07:11 crc kubenswrapper[4802]: E1125 17:07:11.368654 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e244ce8-d122-4a53-bc3e-e42ce7231332" containerName="glance-httpd" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.368677 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e244ce8-d122-4a53-bc3e-e42ce7231332" containerName="glance-httpd" Nov 25 17:07:11 crc kubenswrapper[4802]: E1125 17:07:11.368703 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e244ce8-d122-4a53-bc3e-e42ce7231332" containerName="glance-log" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.368712 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e244ce8-d122-4a53-bc3e-e42ce7231332" containerName="glance-log" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.368891 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e244ce8-d122-4a53-bc3e-e42ce7231332" containerName="glance-log" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.368920 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e244ce8-d122-4a53-bc3e-e42ce7231332" containerName="glance-httpd" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.369860 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.383576 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.420056 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-etc-nvme\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.420096 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-dev\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.420139 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-lib-modules\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.420158 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a987f336-f102-45e2-a131-ca8daa8ffb89-config-data\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.420302 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcvjc\" (UniqueName: \"kubernetes.io/projected/a987f336-f102-45e2-a131-ca8daa8ffb89-kube-api-access-wcvjc\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.420395 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.420487 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.420513 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a987f336-f102-45e2-a131-ca8daa8ffb89-logs\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.420616 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.420649 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a987f336-f102-45e2-a131-ca8daa8ffb89-scripts\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.420745 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a987f336-f102-45e2-a131-ca8daa8ffb89-httpd-run\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.420839 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-run\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.421017 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.421099 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-sys\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.519994 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e244ce8-d122-4a53-bc3e-e42ce7231332" path="/var/lib/kubelet/pods/7e244ce8-d122-4a53-bc3e-e42ce7231332/volumes" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.523418 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-run\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.523496 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.523540 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-sys\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.523599 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-etc-nvme\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.523632 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-dev\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.523664 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-lib-modules\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.523692 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a987f336-f102-45e2-a131-ca8daa8ffb89-config-data\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.523730 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcvjc\" (UniqueName: \"kubernetes.io/projected/a987f336-f102-45e2-a131-ca8daa8ffb89-kube-api-access-wcvjc\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.523760 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.523800 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.523829 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a987f336-f102-45e2-a131-ca8daa8ffb89-logs\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.523864 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.523889 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a987f336-f102-45e2-a131-ca8daa8ffb89-scripts\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.523920 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a987f336-f102-45e2-a131-ca8daa8ffb89-httpd-run\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.524446 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-run\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.524463 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-sys\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.524525 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.524442 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.524588 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-dev\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.524613 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.524677 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-etc-nvme\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.524597 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-lib-modules\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.524709 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.525341 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a987f336-f102-45e2-a131-ca8daa8ffb89-httpd-run\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.525379 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a987f336-f102-45e2-a131-ca8daa8ffb89-logs\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.534071 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a987f336-f102-45e2-a131-ca8daa8ffb89-scripts\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.537868 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a987f336-f102-45e2-a131-ca8daa8ffb89-config-data\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.551821 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcvjc\" (UniqueName: \"kubernetes.io/projected/a987f336-f102-45e2-a131-ca8daa8ffb89-kube-api-access-wcvjc\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.564657 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.576288 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:11 crc kubenswrapper[4802]: I1125 17:07:11.687080 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:12 crc kubenswrapper[4802]: I1125 17:07:12.203416 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 17:07:12 crc kubenswrapper[4802]: W1125 17:07:12.209366 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda987f336_f102_45e2_a131_ca8daa8ffb89.slice/crio-03e9115dd62b56da75ef1d30fd1d43f7301219e0b37a53ef80086e13fce0800c WatchSource:0}: Error finding container 03e9115dd62b56da75ef1d30fd1d43f7301219e0b37a53ef80086e13fce0800c: Status 404 returned error can't find the container with id 03e9115dd62b56da75ef1d30fd1d43f7301219e0b37a53ef80086e13fce0800c Nov 25 17:07:12 crc kubenswrapper[4802]: I1125 17:07:12.300798 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"a987f336-f102-45e2-a131-ca8daa8ffb89","Type":"ContainerStarted","Data":"03e9115dd62b56da75ef1d30fd1d43f7301219e0b37a53ef80086e13fce0800c"} Nov 25 17:07:13 crc kubenswrapper[4802]: I1125 17:07:13.309720 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"a987f336-f102-45e2-a131-ca8daa8ffb89","Type":"ContainerStarted","Data":"4c80c14c72ffe4859658da9a729415d87c1dd0571e0dd8c78f2d7503021180c6"} Nov 25 17:07:13 crc kubenswrapper[4802]: I1125 17:07:13.310319 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"a987f336-f102-45e2-a131-ca8daa8ffb89","Type":"ContainerStarted","Data":"b72932cd979f93ceb1c6b80419ab5d3d58d8b1a9915209bc58a659dbacc29795"} Nov 25 17:07:13 crc kubenswrapper[4802]: I1125 17:07:13.338713 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.338677874 podStartE2EDuration="2.338677874s" podCreationTimestamp="2025-11-25 17:07:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:07:13.336503575 +0000 UTC m=+1216.480850771" watchObservedRunningTime="2025-11-25 17:07:13.338677874 +0000 UTC m=+1216.483025070" Nov 25 17:07:21 crc kubenswrapper[4802]: I1125 17:07:21.688015 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:21 crc kubenswrapper[4802]: I1125 17:07:21.688558 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:21 crc kubenswrapper[4802]: I1125 17:07:21.719198 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:21 crc kubenswrapper[4802]: I1125 17:07:21.753998 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:22 crc kubenswrapper[4802]: I1125 17:07:22.378909 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:22 crc kubenswrapper[4802]: I1125 17:07:22.378998 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:24 crc kubenswrapper[4802]: I1125 17:07:24.370186 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:24 crc kubenswrapper[4802]: I1125 17:07:24.393279 4802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 17:07:25 crc kubenswrapper[4802]: I1125 17:07:25.201208 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:37 crc kubenswrapper[4802]: I1125 17:07:37.991587 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-lspxq"] Nov 25 17:07:37 crc kubenswrapper[4802]: I1125 17:07:37.997352 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-lspxq"] Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.073312 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.073591 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="f2f664f6-2086-40bb-9cfe-0544dfabd571" containerName="glance-log" containerID="cri-o://239b5f614d0559a9c8d73d157796461e709207d744d4d00f57888b782d3f328c" gracePeriod=30 Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.074167 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="f2f664f6-2086-40bb-9cfe-0544dfabd571" containerName="glance-httpd" containerID="cri-o://3208ac757d732c9a0b5234760fd13ad44ecd4525c8b4cd547cf8bfcf6f04cd61" gracePeriod=30 Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.077699 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.077928 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="a987f336-f102-45e2-a131-ca8daa8ffb89" containerName="glance-log" containerID="cri-o://b72932cd979f93ceb1c6b80419ab5d3d58d8b1a9915209bc58a659dbacc29795" gracePeriod=30 Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.077976 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="a987f336-f102-45e2-a131-ca8daa8ffb89" containerName="glance-httpd" containerID="cri-o://4c80c14c72ffe4859658da9a729415d87c1dd0571e0dd8c78f2d7503021180c6" gracePeriod=30 Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.109974 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glancece03-account-delete-ll8qv"] Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.116738 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancece03-account-delete-ll8qv" Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.128552 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancece03-account-delete-ll8qv"] Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.237879 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.238147 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstackclient" podUID="d9d79e72-cdd7-41be-886b-c72137db15bf" containerName="openstackclient" containerID="cri-o://5a791aa521fc8c3c6a0d04b3916489ef736afb9b3a9fcf3c35b6a9c60c7e0e42" gracePeriod=30 Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.253453 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfspw\" (UniqueName: \"kubernetes.io/projected/f5abee75-071c-48fe-8b20-b297da178661-kube-api-access-mfspw\") pod \"glancece03-account-delete-ll8qv\" (UID: \"f5abee75-071c-48fe-8b20-b297da178661\") " pod="glance-kuttl-tests/glancece03-account-delete-ll8qv" Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.253569 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5abee75-071c-48fe-8b20-b297da178661-operator-scripts\") pod \"glancece03-account-delete-ll8qv\" (UID: \"f5abee75-071c-48fe-8b20-b297da178661\") " pod="glance-kuttl-tests/glancece03-account-delete-ll8qv" Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.354588 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfspw\" (UniqueName: \"kubernetes.io/projected/f5abee75-071c-48fe-8b20-b297da178661-kube-api-access-mfspw\") pod \"glancece03-account-delete-ll8qv\" (UID: \"f5abee75-071c-48fe-8b20-b297da178661\") " pod="glance-kuttl-tests/glancece03-account-delete-ll8qv" Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.354670 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5abee75-071c-48fe-8b20-b297da178661-operator-scripts\") pod \"glancece03-account-delete-ll8qv\" (UID: \"f5abee75-071c-48fe-8b20-b297da178661\") " pod="glance-kuttl-tests/glancece03-account-delete-ll8qv" Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.355662 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5abee75-071c-48fe-8b20-b297da178661-operator-scripts\") pod \"glancece03-account-delete-ll8qv\" (UID: \"f5abee75-071c-48fe-8b20-b297da178661\") " pod="glance-kuttl-tests/glancece03-account-delete-ll8qv" Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.375409 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfspw\" (UniqueName: \"kubernetes.io/projected/f5abee75-071c-48fe-8b20-b297da178661-kube-api-access-mfspw\") pod \"glancece03-account-delete-ll8qv\" (UID: \"f5abee75-071c-48fe-8b20-b297da178661\") " pod="glance-kuttl-tests/glancece03-account-delete-ll8qv" Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.433021 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancece03-account-delete-ll8qv" Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.508652 4802 generic.go:334] "Generic (PLEG): container finished" podID="d9d79e72-cdd7-41be-886b-c72137db15bf" containerID="5a791aa521fc8c3c6a0d04b3916489ef736afb9b3a9fcf3c35b6a9c60c7e0e42" exitCode=143 Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.508879 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"d9d79e72-cdd7-41be-886b-c72137db15bf","Type":"ContainerDied","Data":"5a791aa521fc8c3c6a0d04b3916489ef736afb9b3a9fcf3c35b6a9c60c7e0e42"} Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.512584 4802 generic.go:334] "Generic (PLEG): container finished" podID="f2f664f6-2086-40bb-9cfe-0544dfabd571" containerID="239b5f614d0559a9c8d73d157796461e709207d744d4d00f57888b782d3f328c" exitCode=143 Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.512628 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"f2f664f6-2086-40bb-9cfe-0544dfabd571","Type":"ContainerDied","Data":"239b5f614d0559a9c8d73d157796461e709207d744d4d00f57888b782d3f328c"} Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.520030 4802 generic.go:334] "Generic (PLEG): container finished" podID="a987f336-f102-45e2-a131-ca8daa8ffb89" containerID="b72932cd979f93ceb1c6b80419ab5d3d58d8b1a9915209bc58a659dbacc29795" exitCode=143 Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.520059 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"a987f336-f102-45e2-a131-ca8daa8ffb89","Type":"ContainerDied","Data":"b72932cd979f93ceb1c6b80419ab5d3d58d8b1a9915209bc58a659dbacc29795"} Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.589946 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.658930 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-config-secret\") pod \"d9d79e72-cdd7-41be-886b-c72137db15bf\" (UID: \"d9d79e72-cdd7-41be-886b-c72137db15bf\") " Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.659027 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-scripts\") pod \"d9d79e72-cdd7-41be-886b-c72137db15bf\" (UID: \"d9d79e72-cdd7-41be-886b-c72137db15bf\") " Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.659074 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6rf5\" (UniqueName: \"kubernetes.io/projected/d9d79e72-cdd7-41be-886b-c72137db15bf-kube-api-access-t6rf5\") pod \"d9d79e72-cdd7-41be-886b-c72137db15bf\" (UID: \"d9d79e72-cdd7-41be-886b-c72137db15bf\") " Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.659202 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-config\") pod \"d9d79e72-cdd7-41be-886b-c72137db15bf\" (UID: \"d9d79e72-cdd7-41be-886b-c72137db15bf\") " Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.659941 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-scripts" (OuterVolumeSpecName: "openstack-scripts") pod "d9d79e72-cdd7-41be-886b-c72137db15bf" (UID: "d9d79e72-cdd7-41be-886b-c72137db15bf"). InnerVolumeSpecName "openstack-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.667333 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9d79e72-cdd7-41be-886b-c72137db15bf-kube-api-access-t6rf5" (OuterVolumeSpecName: "kube-api-access-t6rf5") pod "d9d79e72-cdd7-41be-886b-c72137db15bf" (UID: "d9d79e72-cdd7-41be-886b-c72137db15bf"). InnerVolumeSpecName "kube-api-access-t6rf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:07:38 crc kubenswrapper[4802]: E1125 17:07:38.675641 4802 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-config podName:d9d79e72-cdd7-41be-886b-c72137db15bf nodeName:}" failed. No retries permitted until 2025-11-25 17:07:39.175611942 +0000 UTC m=+1242.319959128 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-config") pod "d9d79e72-cdd7-41be-886b-c72137db15bf" (UID: "d9d79e72-cdd7-41be-886b-c72137db15bf") : error deleting /var/lib/kubelet/pods/d9d79e72-cdd7-41be-886b-c72137db15bf/volume-subpaths: remove /var/lib/kubelet/pods/d9d79e72-cdd7-41be-886b-c72137db15bf/volume-subpaths: no such file or directory Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.679500 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "d9d79e72-cdd7-41be-886b-c72137db15bf" (UID: "d9d79e72-cdd7-41be-886b-c72137db15bf"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.760889 4802 reconciler_common.go:293] "Volume detached for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.760947 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6rf5\" (UniqueName: \"kubernetes.io/projected/d9d79e72-cdd7-41be-886b-c72137db15bf-kube-api-access-t6rf5\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.760969 4802 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:38 crc kubenswrapper[4802]: I1125 17:07:38.899520 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancece03-account-delete-ll8qv"] Nov 25 17:07:38 crc kubenswrapper[4802]: W1125 17:07:38.906165 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5abee75_071c_48fe_8b20_b297da178661.slice/crio-09e88d9581b09ed3a53f292759644e6b0c6f47421ef421a26c1ba130339b3d9d WatchSource:0}: Error finding container 09e88d9581b09ed3a53f292759644e6b0c6f47421ef421a26c1ba130339b3d9d: Status 404 returned error can't find the container with id 09e88d9581b09ed3a53f292759644e6b0c6f47421ef421a26c1ba130339b3d9d Nov 25 17:07:39 crc kubenswrapper[4802]: I1125 17:07:39.266992 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-config\") pod \"d9d79e72-cdd7-41be-886b-c72137db15bf\" (UID: \"d9d79e72-cdd7-41be-886b-c72137db15bf\") " Nov 25 17:07:39 crc kubenswrapper[4802]: I1125 17:07:39.268107 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "d9d79e72-cdd7-41be-886b-c72137db15bf" (UID: "d9d79e72-cdd7-41be-886b-c72137db15bf"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:07:39 crc kubenswrapper[4802]: I1125 17:07:39.369603 4802 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d9d79e72-cdd7-41be-886b-c72137db15bf-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:39 crc kubenswrapper[4802]: I1125 17:07:39.522050 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ab3f03b-f8da-4082-8ece-aa2d9375accf" path="/var/lib/kubelet/pods/5ab3f03b-f8da-4082-8ece-aa2d9375accf/volumes" Nov 25 17:07:39 crc kubenswrapper[4802]: I1125 17:07:39.530509 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 17:07:39 crc kubenswrapper[4802]: I1125 17:07:39.530498 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"d9d79e72-cdd7-41be-886b-c72137db15bf","Type":"ContainerDied","Data":"2d607d882fd73b4e4a8f4ae6ab876f25e02524a51094516986f2c9b75ccd4b15"} Nov 25 17:07:39 crc kubenswrapper[4802]: I1125 17:07:39.530677 4802 scope.go:117] "RemoveContainer" containerID="5a791aa521fc8c3c6a0d04b3916489ef736afb9b3a9fcf3c35b6a9c60c7e0e42" Nov 25 17:07:39 crc kubenswrapper[4802]: I1125 17:07:39.532935 4802 generic.go:334] "Generic (PLEG): container finished" podID="f5abee75-071c-48fe-8b20-b297da178661" containerID="a80bf8be329c35e0a009bd66ada0f49aeb9fd69ec21cbca077fd1c16637f88ba" exitCode=0 Nov 25 17:07:39 crc kubenswrapper[4802]: I1125 17:07:39.532987 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancece03-account-delete-ll8qv" event={"ID":"f5abee75-071c-48fe-8b20-b297da178661","Type":"ContainerDied","Data":"a80bf8be329c35e0a009bd66ada0f49aeb9fd69ec21cbca077fd1c16637f88ba"} Nov 25 17:07:39 crc kubenswrapper[4802]: I1125 17:07:39.533016 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancece03-account-delete-ll8qv" event={"ID":"f5abee75-071c-48fe-8b20-b297da178661","Type":"ContainerStarted","Data":"09e88d9581b09ed3a53f292759644e6b0c6f47421ef421a26c1ba130339b3d9d"} Nov 25 17:07:39 crc kubenswrapper[4802]: I1125 17:07:39.564545 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 17:07:39 crc kubenswrapper[4802]: I1125 17:07:39.571315 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 17:07:40 crc kubenswrapper[4802]: I1125 17:07:40.828057 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancece03-account-delete-ll8qv" Nov 25 17:07:40 crc kubenswrapper[4802]: I1125 17:07:40.892986 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfspw\" (UniqueName: \"kubernetes.io/projected/f5abee75-071c-48fe-8b20-b297da178661-kube-api-access-mfspw\") pod \"f5abee75-071c-48fe-8b20-b297da178661\" (UID: \"f5abee75-071c-48fe-8b20-b297da178661\") " Nov 25 17:07:40 crc kubenswrapper[4802]: I1125 17:07:40.893071 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5abee75-071c-48fe-8b20-b297da178661-operator-scripts\") pod \"f5abee75-071c-48fe-8b20-b297da178661\" (UID: \"f5abee75-071c-48fe-8b20-b297da178661\") " Nov 25 17:07:40 crc kubenswrapper[4802]: I1125 17:07:40.893825 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5abee75-071c-48fe-8b20-b297da178661-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f5abee75-071c-48fe-8b20-b297da178661" (UID: "f5abee75-071c-48fe-8b20-b297da178661"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:07:40 crc kubenswrapper[4802]: I1125 17:07:40.897790 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5abee75-071c-48fe-8b20-b297da178661-kube-api-access-mfspw" (OuterVolumeSpecName: "kube-api-access-mfspw") pod "f5abee75-071c-48fe-8b20-b297da178661" (UID: "f5abee75-071c-48fe-8b20-b297da178661"). InnerVolumeSpecName "kube-api-access-mfspw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:07:40 crc kubenswrapper[4802]: I1125 17:07:40.994860 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5abee75-071c-48fe-8b20-b297da178661-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:40 crc kubenswrapper[4802]: I1125 17:07:40.994908 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfspw\" (UniqueName: \"kubernetes.io/projected/f5abee75-071c-48fe-8b20-b297da178661-kube-api-access-mfspw\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.515155 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9d79e72-cdd7-41be-886b-c72137db15bf" path="/var/lib/kubelet/pods/d9d79e72-cdd7-41be-886b-c72137db15bf/volumes" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.557960 4802 generic.go:334] "Generic (PLEG): container finished" podID="a987f336-f102-45e2-a131-ca8daa8ffb89" containerID="4c80c14c72ffe4859658da9a729415d87c1dd0571e0dd8c78f2d7503021180c6" exitCode=0 Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.558045 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"a987f336-f102-45e2-a131-ca8daa8ffb89","Type":"ContainerDied","Data":"4c80c14c72ffe4859658da9a729415d87c1dd0571e0dd8c78f2d7503021180c6"} Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.560135 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancece03-account-delete-ll8qv" event={"ID":"f5abee75-071c-48fe-8b20-b297da178661","Type":"ContainerDied","Data":"09e88d9581b09ed3a53f292759644e6b0c6f47421ef421a26c1ba130339b3d9d"} Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.560156 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancece03-account-delete-ll8qv" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.560164 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09e88d9581b09ed3a53f292759644e6b0c6f47421ef421a26c1ba130339b3d9d" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.564205 4802 generic.go:334] "Generic (PLEG): container finished" podID="f2f664f6-2086-40bb-9cfe-0544dfabd571" containerID="3208ac757d732c9a0b5234760fd13ad44ecd4525c8b4cd547cf8bfcf6f04cd61" exitCode=0 Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.564247 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"f2f664f6-2086-40bb-9cfe-0544dfabd571","Type":"ContainerDied","Data":"3208ac757d732c9a0b5234760fd13ad44ecd4525c8b4cd547cf8bfcf6f04cd61"} Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.595967 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.633493 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.704427 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"a987f336-f102-45e2-a131-ca8daa8ffb89\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.704500 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vp99\" (UniqueName: \"kubernetes.io/projected/f2f664f6-2086-40bb-9cfe-0544dfabd571-kube-api-access-2vp99\") pod \"f2f664f6-2086-40bb-9cfe-0544dfabd571\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.704553 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2f664f6-2086-40bb-9cfe-0544dfabd571-logs\") pod \"f2f664f6-2086-40bb-9cfe-0544dfabd571\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.704583 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a987f336-f102-45e2-a131-ca8daa8ffb89-scripts\") pod \"a987f336-f102-45e2-a131-ca8daa8ffb89\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.704625 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a987f336-f102-45e2-a131-ca8daa8ffb89-config-data\") pod \"a987f336-f102-45e2-a131-ca8daa8ffb89\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.704657 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a987f336-f102-45e2-a131-ca8daa8ffb89-logs\") pod \"a987f336-f102-45e2-a131-ca8daa8ffb89\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.704704 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-dev\") pod \"f2f664f6-2086-40bb-9cfe-0544dfabd571\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.704721 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2f664f6-2086-40bb-9cfe-0544dfabd571-config-data\") pod \"f2f664f6-2086-40bb-9cfe-0544dfabd571\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.704747 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2f664f6-2086-40bb-9cfe-0544dfabd571-httpd-run\") pod \"f2f664f6-2086-40bb-9cfe-0544dfabd571\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.704789 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-etc-iscsi\") pod \"a987f336-f102-45e2-a131-ca8daa8ffb89\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.704817 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-lib-modules\") pod \"f2f664f6-2086-40bb-9cfe-0544dfabd571\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.704832 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-lib-modules\") pod \"a987f336-f102-45e2-a131-ca8daa8ffb89\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.704844 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-var-locks-brick\") pod \"a987f336-f102-45e2-a131-ca8daa8ffb89\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.704898 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-etc-nvme\") pod \"a987f336-f102-45e2-a131-ca8daa8ffb89\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.704973 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"a987f336-f102-45e2-a131-ca8daa8ffb89\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.705034 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-run\") pod \"a987f336-f102-45e2-a131-ca8daa8ffb89\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.705096 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-dev\") pod \"a987f336-f102-45e2-a131-ca8daa8ffb89\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.705153 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"f2f664f6-2086-40bb-9cfe-0544dfabd571\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.705173 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-run\") pod \"f2f664f6-2086-40bb-9cfe-0544dfabd571\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.705195 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-sys\") pod \"a987f336-f102-45e2-a131-ca8daa8ffb89\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.705212 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-var-locks-brick\") pod \"f2f664f6-2086-40bb-9cfe-0544dfabd571\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.705293 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a987f336-f102-45e2-a131-ca8daa8ffb89-httpd-run\") pod \"a987f336-f102-45e2-a131-ca8daa8ffb89\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.705311 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-sys\") pod \"f2f664f6-2086-40bb-9cfe-0544dfabd571\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.705328 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcvjc\" (UniqueName: \"kubernetes.io/projected/a987f336-f102-45e2-a131-ca8daa8ffb89-kube-api-access-wcvjc\") pod \"a987f336-f102-45e2-a131-ca8daa8ffb89\" (UID: \"a987f336-f102-45e2-a131-ca8daa8ffb89\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.705356 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-etc-iscsi\") pod \"f2f664f6-2086-40bb-9cfe-0544dfabd571\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.705371 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"f2f664f6-2086-40bb-9cfe-0544dfabd571\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.705387 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2f664f6-2086-40bb-9cfe-0544dfabd571-scripts\") pod \"f2f664f6-2086-40bb-9cfe-0544dfabd571\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.705402 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-etc-nvme\") pod \"f2f664f6-2086-40bb-9cfe-0544dfabd571\" (UID: \"f2f664f6-2086-40bb-9cfe-0544dfabd571\") " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.706526 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2f664f6-2086-40bb-9cfe-0544dfabd571-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f2f664f6-2086-40bb-9cfe-0544dfabd571" (UID: "f2f664f6-2086-40bb-9cfe-0544dfabd571"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.706569 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "a987f336-f102-45e2-a131-ca8daa8ffb89" (UID: "a987f336-f102-45e2-a131-ca8daa8ffb89"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.706589 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "f2f664f6-2086-40bb-9cfe-0544dfabd571" (UID: "f2f664f6-2086-40bb-9cfe-0544dfabd571"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.706606 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "a987f336-f102-45e2-a131-ca8daa8ffb89" (UID: "a987f336-f102-45e2-a131-ca8daa8ffb89"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.706622 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "a987f336-f102-45e2-a131-ca8daa8ffb89" (UID: "a987f336-f102-45e2-a131-ca8daa8ffb89"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.706638 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "a987f336-f102-45e2-a131-ca8daa8ffb89" (UID: "a987f336-f102-45e2-a131-ca8daa8ffb89"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.707235 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a987f336-f102-45e2-a131-ca8daa8ffb89-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a987f336-f102-45e2-a131-ca8daa8ffb89" (UID: "a987f336-f102-45e2-a131-ca8daa8ffb89"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.707266 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-run" (OuterVolumeSpecName: "run") pod "a987f336-f102-45e2-a131-ca8daa8ffb89" (UID: "a987f336-f102-45e2-a131-ca8daa8ffb89"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.707284 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-dev" (OuterVolumeSpecName: "dev") pod "a987f336-f102-45e2-a131-ca8daa8ffb89" (UID: "a987f336-f102-45e2-a131-ca8daa8ffb89"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.707452 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-sys" (OuterVolumeSpecName: "sys") pod "a987f336-f102-45e2-a131-ca8daa8ffb89" (UID: "a987f336-f102-45e2-a131-ca8daa8ffb89"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.707533 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-run" (OuterVolumeSpecName: "run") pod "f2f664f6-2086-40bb-9cfe-0544dfabd571" (UID: "f2f664f6-2086-40bb-9cfe-0544dfabd571"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.711897 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a987f336-f102-45e2-a131-ca8daa8ffb89-scripts" (OuterVolumeSpecName: "scripts") pod "a987f336-f102-45e2-a131-ca8daa8ffb89" (UID: "a987f336-f102-45e2-a131-ca8daa8ffb89"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.712008 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "f2f664f6-2086-40bb-9cfe-0544dfabd571" (UID: "f2f664f6-2086-40bb-9cfe-0544dfabd571"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.712048 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "f2f664f6-2086-40bb-9cfe-0544dfabd571" (UID: "f2f664f6-2086-40bb-9cfe-0544dfabd571"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.712143 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "a987f336-f102-45e2-a131-ca8daa8ffb89" (UID: "a987f336-f102-45e2-a131-ca8daa8ffb89"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.712351 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2f664f6-2086-40bb-9cfe-0544dfabd571-logs" (OuterVolumeSpecName: "logs") pod "f2f664f6-2086-40bb-9cfe-0544dfabd571" (UID: "f2f664f6-2086-40bb-9cfe-0544dfabd571"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.712398 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "f2f664f6-2086-40bb-9cfe-0544dfabd571" (UID: "f2f664f6-2086-40bb-9cfe-0544dfabd571"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.712972 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a987f336-f102-45e2-a131-ca8daa8ffb89-logs" (OuterVolumeSpecName: "logs") pod "a987f336-f102-45e2-a131-ca8daa8ffb89" (UID: "a987f336-f102-45e2-a131-ca8daa8ffb89"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.714262 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2f664f6-2086-40bb-9cfe-0544dfabd571-kube-api-access-2vp99" (OuterVolumeSpecName: "kube-api-access-2vp99") pod "f2f664f6-2086-40bb-9cfe-0544dfabd571" (UID: "f2f664f6-2086-40bb-9cfe-0544dfabd571"). InnerVolumeSpecName "kube-api-access-2vp99". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.714319 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-dev" (OuterVolumeSpecName: "dev") pod "f2f664f6-2086-40bb-9cfe-0544dfabd571" (UID: "f2f664f6-2086-40bb-9cfe-0544dfabd571"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.714328 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2f664f6-2086-40bb-9cfe-0544dfabd571-scripts" (OuterVolumeSpecName: "scripts") pod "f2f664f6-2086-40bb-9cfe-0544dfabd571" (UID: "f2f664f6-2086-40bb-9cfe-0544dfabd571"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.714361 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "f2f664f6-2086-40bb-9cfe-0544dfabd571" (UID: "f2f664f6-2086-40bb-9cfe-0544dfabd571"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.714380 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-sys" (OuterVolumeSpecName: "sys") pod "f2f664f6-2086-40bb-9cfe-0544dfabd571" (UID: "f2f664f6-2086-40bb-9cfe-0544dfabd571"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.715223 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance-cache") pod "f2f664f6-2086-40bb-9cfe-0544dfabd571" (UID: "f2f664f6-2086-40bb-9cfe-0544dfabd571"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.716816 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "a987f336-f102-45e2-a131-ca8daa8ffb89" (UID: "a987f336-f102-45e2-a131-ca8daa8ffb89"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.717742 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a987f336-f102-45e2-a131-ca8daa8ffb89-kube-api-access-wcvjc" (OuterVolumeSpecName: "kube-api-access-wcvjc") pod "a987f336-f102-45e2-a131-ca8daa8ffb89" (UID: "a987f336-f102-45e2-a131-ca8daa8ffb89"). InnerVolumeSpecName "kube-api-access-wcvjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.752550 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2f664f6-2086-40bb-9cfe-0544dfabd571-config-data" (OuterVolumeSpecName: "config-data") pod "f2f664f6-2086-40bb-9cfe-0544dfabd571" (UID: "f2f664f6-2086-40bb-9cfe-0544dfabd571"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.754592 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a987f336-f102-45e2-a131-ca8daa8ffb89-config-data" (OuterVolumeSpecName: "config-data") pod "a987f336-f102-45e2-a131-ca8daa8ffb89" (UID: "a987f336-f102-45e2-a131-ca8daa8ffb89"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.806979 4802 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-dev\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807018 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2f664f6-2086-40bb-9cfe-0544dfabd571-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807034 4802 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2f664f6-2086-40bb-9cfe-0544dfabd571-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807046 4802 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807061 4802 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807076 4802 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807088 4802 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807100 4802 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807111 4802 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807167 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807181 4802 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-dev\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807194 4802 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807211 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807223 4802 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a987f336-f102-45e2-a131-ca8daa8ffb89-sys\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807234 4802 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807247 4802 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a987f336-f102-45e2-a131-ca8daa8ffb89-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807260 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcvjc\" (UniqueName: \"kubernetes.io/projected/a987f336-f102-45e2-a131-ca8daa8ffb89-kube-api-access-wcvjc\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807271 4802 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-sys\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807282 4802 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807299 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807311 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2f664f6-2086-40bb-9cfe-0544dfabd571-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807322 4802 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f2f664f6-2086-40bb-9cfe-0544dfabd571-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807337 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807349 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vp99\" (UniqueName: \"kubernetes.io/projected/f2f664f6-2086-40bb-9cfe-0544dfabd571-kube-api-access-2vp99\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807360 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2f664f6-2086-40bb-9cfe-0544dfabd571-logs\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807371 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a987f336-f102-45e2-a131-ca8daa8ffb89-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807382 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a987f336-f102-45e2-a131-ca8daa8ffb89-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.807395 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a987f336-f102-45e2-a131-ca8daa8ffb89-logs\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.823062 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.823804 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.823866 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.824027 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.908418 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.908453 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.908465 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:41 crc kubenswrapper[4802]: I1125 17:07:41.908478 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:42 crc kubenswrapper[4802]: I1125 17:07:42.574332 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"f2f664f6-2086-40bb-9cfe-0544dfabd571","Type":"ContainerDied","Data":"2eeebce12b45fe68cf89f29b7bb76923c043a720b6e2c914589f09b3f073849c"} Nov 25 17:07:42 crc kubenswrapper[4802]: I1125 17:07:42.574372 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 17:07:42 crc kubenswrapper[4802]: I1125 17:07:42.574675 4802 scope.go:117] "RemoveContainer" containerID="3208ac757d732c9a0b5234760fd13ad44ecd4525c8b4cd547cf8bfcf6f04cd61" Nov 25 17:07:42 crc kubenswrapper[4802]: I1125 17:07:42.577382 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"a987f336-f102-45e2-a131-ca8daa8ffb89","Type":"ContainerDied","Data":"03e9115dd62b56da75ef1d30fd1d43f7301219e0b37a53ef80086e13fce0800c"} Nov 25 17:07:42 crc kubenswrapper[4802]: I1125 17:07:42.577464 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:42 crc kubenswrapper[4802]: I1125 17:07:42.601294 4802 scope.go:117] "RemoveContainer" containerID="239b5f614d0559a9c8d73d157796461e709207d744d4d00f57888b782d3f328c" Nov 25 17:07:42 crc kubenswrapper[4802]: I1125 17:07:42.627151 4802 scope.go:117] "RemoveContainer" containerID="4c80c14c72ffe4859658da9a729415d87c1dd0571e0dd8c78f2d7503021180c6" Nov 25 17:07:42 crc kubenswrapper[4802]: I1125 17:07:42.627490 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 17:07:42 crc kubenswrapper[4802]: I1125 17:07:42.637272 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 17:07:42 crc kubenswrapper[4802]: I1125 17:07:42.642606 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 17:07:42 crc kubenswrapper[4802]: I1125 17:07:42.648408 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 17:07:42 crc kubenswrapper[4802]: I1125 17:07:42.649836 4802 scope.go:117] "RemoveContainer" containerID="b72932cd979f93ceb1c6b80419ab5d3d58d8b1a9915209bc58a659dbacc29795" Nov 25 17:07:43 crc kubenswrapper[4802]: I1125 17:07:43.107994 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-fxkc8"] Nov 25 17:07:43 crc kubenswrapper[4802]: I1125 17:07:43.114841 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-fxkc8"] Nov 25 17:07:43 crc kubenswrapper[4802]: I1125 17:07:43.123667 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glancece03-account-delete-ll8qv"] Nov 25 17:07:43 crc kubenswrapper[4802]: I1125 17:07:43.137695 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glancece03-account-delete-ll8qv"] Nov 25 17:07:43 crc kubenswrapper[4802]: I1125 17:07:43.144755 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-ce03-account-create-update-ds8w8"] Nov 25 17:07:43 crc kubenswrapper[4802]: I1125 17:07:43.151425 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-ce03-account-create-update-ds8w8"] Nov 25 17:07:43 crc kubenswrapper[4802]: I1125 17:07:43.515057 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44f61cc5-32b0-4f09-919a-c1d010c2ab28" path="/var/lib/kubelet/pods/44f61cc5-32b0-4f09-919a-c1d010c2ab28/volumes" Nov 25 17:07:43 crc kubenswrapper[4802]: I1125 17:07:43.515890 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a987f336-f102-45e2-a131-ca8daa8ffb89" path="/var/lib/kubelet/pods/a987f336-f102-45e2-a131-ca8daa8ffb89/volumes" Nov 25 17:07:43 crc kubenswrapper[4802]: I1125 17:07:43.516486 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae47311c-06d0-4f16-a989-d0a60f7d7faa" path="/var/lib/kubelet/pods/ae47311c-06d0-4f16-a989-d0a60f7d7faa/volumes" Nov 25 17:07:43 crc kubenswrapper[4802]: I1125 17:07:43.517061 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2f664f6-2086-40bb-9cfe-0544dfabd571" path="/var/lib/kubelet/pods/f2f664f6-2086-40bb-9cfe-0544dfabd571/volumes" Nov 25 17:07:43 crc kubenswrapper[4802]: I1125 17:07:43.518025 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5abee75-071c-48fe-8b20-b297da178661" path="/var/lib/kubelet/pods/f5abee75-071c-48fe-8b20-b297da178661/volumes" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.853222 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-8sxkq"] Nov 25 17:07:44 crc kubenswrapper[4802]: E1125 17:07:44.853672 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2f664f6-2086-40bb-9cfe-0544dfabd571" containerName="glance-log" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.853684 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2f664f6-2086-40bb-9cfe-0544dfabd571" containerName="glance-log" Nov 25 17:07:44 crc kubenswrapper[4802]: E1125 17:07:44.853700 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2f664f6-2086-40bb-9cfe-0544dfabd571" containerName="glance-httpd" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.853705 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2f664f6-2086-40bb-9cfe-0544dfabd571" containerName="glance-httpd" Nov 25 17:07:44 crc kubenswrapper[4802]: E1125 17:07:44.853716 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d79e72-cdd7-41be-886b-c72137db15bf" containerName="openstackclient" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.853722 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d79e72-cdd7-41be-886b-c72137db15bf" containerName="openstackclient" Nov 25 17:07:44 crc kubenswrapper[4802]: E1125 17:07:44.853737 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a987f336-f102-45e2-a131-ca8daa8ffb89" containerName="glance-log" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.853743 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a987f336-f102-45e2-a131-ca8daa8ffb89" containerName="glance-log" Nov 25 17:07:44 crc kubenswrapper[4802]: E1125 17:07:44.853757 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a987f336-f102-45e2-a131-ca8daa8ffb89" containerName="glance-httpd" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.853762 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a987f336-f102-45e2-a131-ca8daa8ffb89" containerName="glance-httpd" Nov 25 17:07:44 crc kubenswrapper[4802]: E1125 17:07:44.853771 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5abee75-071c-48fe-8b20-b297da178661" containerName="mariadb-account-delete" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.853777 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5abee75-071c-48fe-8b20-b297da178661" containerName="mariadb-account-delete" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.853887 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5abee75-071c-48fe-8b20-b297da178661" containerName="mariadb-account-delete" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.853903 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a987f336-f102-45e2-a131-ca8daa8ffb89" containerName="glance-log" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.853913 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2f664f6-2086-40bb-9cfe-0544dfabd571" containerName="glance-httpd" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.853921 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a987f336-f102-45e2-a131-ca8daa8ffb89" containerName="glance-httpd" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.853929 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2f664f6-2086-40bb-9cfe-0544dfabd571" containerName="glance-log" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.853935 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9d79e72-cdd7-41be-886b-c72137db15bf" containerName="openstackclient" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.854364 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-8sxkq" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.861159 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-51cf-account-create-update-lhnnv"] Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.862135 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-51cf-account-create-update-lhnnv" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.863449 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.865666 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-8sxkq"] Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.874629 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-51cf-account-create-update-lhnnv"] Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.955017 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dbfda435-8e90-4baf-ba54-45646e1ca5c9-operator-scripts\") pod \"glance-db-create-8sxkq\" (UID: \"dbfda435-8e90-4baf-ba54-45646e1ca5c9\") " pod="glance-kuttl-tests/glance-db-create-8sxkq" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.955085 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e20085a-408e-42db-b120-bfb28eb25b42-operator-scripts\") pod \"glance-51cf-account-create-update-lhnnv\" (UID: \"2e20085a-408e-42db-b120-bfb28eb25b42\") " pod="glance-kuttl-tests/glance-51cf-account-create-update-lhnnv" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.955143 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdmcl\" (UniqueName: \"kubernetes.io/projected/dbfda435-8e90-4baf-ba54-45646e1ca5c9-kube-api-access-xdmcl\") pod \"glance-db-create-8sxkq\" (UID: \"dbfda435-8e90-4baf-ba54-45646e1ca5c9\") " pod="glance-kuttl-tests/glance-db-create-8sxkq" Nov 25 17:07:44 crc kubenswrapper[4802]: I1125 17:07:44.955200 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxn9f\" (UniqueName: \"kubernetes.io/projected/2e20085a-408e-42db-b120-bfb28eb25b42-kube-api-access-nxn9f\") pod \"glance-51cf-account-create-update-lhnnv\" (UID: \"2e20085a-408e-42db-b120-bfb28eb25b42\") " pod="glance-kuttl-tests/glance-51cf-account-create-update-lhnnv" Nov 25 17:07:45 crc kubenswrapper[4802]: I1125 17:07:45.057018 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e20085a-408e-42db-b120-bfb28eb25b42-operator-scripts\") pod \"glance-51cf-account-create-update-lhnnv\" (UID: \"2e20085a-408e-42db-b120-bfb28eb25b42\") " pod="glance-kuttl-tests/glance-51cf-account-create-update-lhnnv" Nov 25 17:07:45 crc kubenswrapper[4802]: I1125 17:07:45.057076 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdmcl\" (UniqueName: \"kubernetes.io/projected/dbfda435-8e90-4baf-ba54-45646e1ca5c9-kube-api-access-xdmcl\") pod \"glance-db-create-8sxkq\" (UID: \"dbfda435-8e90-4baf-ba54-45646e1ca5c9\") " pod="glance-kuttl-tests/glance-db-create-8sxkq" Nov 25 17:07:45 crc kubenswrapper[4802]: I1125 17:07:45.057165 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxn9f\" (UniqueName: \"kubernetes.io/projected/2e20085a-408e-42db-b120-bfb28eb25b42-kube-api-access-nxn9f\") pod \"glance-51cf-account-create-update-lhnnv\" (UID: \"2e20085a-408e-42db-b120-bfb28eb25b42\") " pod="glance-kuttl-tests/glance-51cf-account-create-update-lhnnv" Nov 25 17:07:45 crc kubenswrapper[4802]: I1125 17:07:45.057228 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dbfda435-8e90-4baf-ba54-45646e1ca5c9-operator-scripts\") pod \"glance-db-create-8sxkq\" (UID: \"dbfda435-8e90-4baf-ba54-45646e1ca5c9\") " pod="glance-kuttl-tests/glance-db-create-8sxkq" Nov 25 17:07:45 crc kubenswrapper[4802]: I1125 17:07:45.057893 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e20085a-408e-42db-b120-bfb28eb25b42-operator-scripts\") pod \"glance-51cf-account-create-update-lhnnv\" (UID: \"2e20085a-408e-42db-b120-bfb28eb25b42\") " pod="glance-kuttl-tests/glance-51cf-account-create-update-lhnnv" Nov 25 17:07:45 crc kubenswrapper[4802]: I1125 17:07:45.058094 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dbfda435-8e90-4baf-ba54-45646e1ca5c9-operator-scripts\") pod \"glance-db-create-8sxkq\" (UID: \"dbfda435-8e90-4baf-ba54-45646e1ca5c9\") " pod="glance-kuttl-tests/glance-db-create-8sxkq" Nov 25 17:07:45 crc kubenswrapper[4802]: I1125 17:07:45.077802 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxn9f\" (UniqueName: \"kubernetes.io/projected/2e20085a-408e-42db-b120-bfb28eb25b42-kube-api-access-nxn9f\") pod \"glance-51cf-account-create-update-lhnnv\" (UID: \"2e20085a-408e-42db-b120-bfb28eb25b42\") " pod="glance-kuttl-tests/glance-51cf-account-create-update-lhnnv" Nov 25 17:07:45 crc kubenswrapper[4802]: I1125 17:07:45.084175 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdmcl\" (UniqueName: \"kubernetes.io/projected/dbfda435-8e90-4baf-ba54-45646e1ca5c9-kube-api-access-xdmcl\") pod \"glance-db-create-8sxkq\" (UID: \"dbfda435-8e90-4baf-ba54-45646e1ca5c9\") " pod="glance-kuttl-tests/glance-db-create-8sxkq" Nov 25 17:07:45 crc kubenswrapper[4802]: I1125 17:07:45.177765 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-8sxkq" Nov 25 17:07:45 crc kubenswrapper[4802]: I1125 17:07:45.183795 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-51cf-account-create-update-lhnnv" Nov 25 17:07:45 crc kubenswrapper[4802]: I1125 17:07:45.448317 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-51cf-account-create-update-lhnnv"] Nov 25 17:07:45 crc kubenswrapper[4802]: I1125 17:07:45.603194 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-8sxkq"] Nov 25 17:07:45 crc kubenswrapper[4802]: I1125 17:07:45.604945 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-51cf-account-create-update-lhnnv" event={"ID":"2e20085a-408e-42db-b120-bfb28eb25b42","Type":"ContainerStarted","Data":"02035075245db14bd28a5710c1fe387e94851d9b3b97d4562711b16f7d2ad92c"} Nov 25 17:07:45 crc kubenswrapper[4802]: W1125 17:07:45.613819 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbfda435_8e90_4baf_ba54_45646e1ca5c9.slice/crio-acb5ac77ea394d7aa475a8647413e98c2cf693934143fc877c34b5c713f61264 WatchSource:0}: Error finding container acb5ac77ea394d7aa475a8647413e98c2cf693934143fc877c34b5c713f61264: Status 404 returned error can't find the container with id acb5ac77ea394d7aa475a8647413e98c2cf693934143fc877c34b5c713f61264 Nov 25 17:07:46 crc kubenswrapper[4802]: I1125 17:07:46.623903 4802 generic.go:334] "Generic (PLEG): container finished" podID="dbfda435-8e90-4baf-ba54-45646e1ca5c9" containerID="9c3ba8f730ea71ef73f67284f774443dfb7dd6eea6ae7fbd8b2f4c97b17fac29" exitCode=0 Nov 25 17:07:46 crc kubenswrapper[4802]: I1125 17:07:46.623995 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-8sxkq" event={"ID":"dbfda435-8e90-4baf-ba54-45646e1ca5c9","Type":"ContainerDied","Data":"9c3ba8f730ea71ef73f67284f774443dfb7dd6eea6ae7fbd8b2f4c97b17fac29"} Nov 25 17:07:46 crc kubenswrapper[4802]: I1125 17:07:46.624029 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-8sxkq" event={"ID":"dbfda435-8e90-4baf-ba54-45646e1ca5c9","Type":"ContainerStarted","Data":"acb5ac77ea394d7aa475a8647413e98c2cf693934143fc877c34b5c713f61264"} Nov 25 17:07:46 crc kubenswrapper[4802]: I1125 17:07:46.626014 4802 generic.go:334] "Generic (PLEG): container finished" podID="2e20085a-408e-42db-b120-bfb28eb25b42" containerID="07fe37447f8badd570e3180f6c54c100c16950d5cf9975117ee0cac3a59616bf" exitCode=0 Nov 25 17:07:46 crc kubenswrapper[4802]: I1125 17:07:46.626072 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-51cf-account-create-update-lhnnv" event={"ID":"2e20085a-408e-42db-b120-bfb28eb25b42","Type":"ContainerDied","Data":"07fe37447f8badd570e3180f6c54c100c16950d5cf9975117ee0cac3a59616bf"} Nov 25 17:07:47 crc kubenswrapper[4802]: I1125 17:07:47.932389 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-8sxkq" Nov 25 17:07:47 crc kubenswrapper[4802]: I1125 17:07:47.996800 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dbfda435-8e90-4baf-ba54-45646e1ca5c9-operator-scripts\") pod \"dbfda435-8e90-4baf-ba54-45646e1ca5c9\" (UID: \"dbfda435-8e90-4baf-ba54-45646e1ca5c9\") " Nov 25 17:07:47 crc kubenswrapper[4802]: I1125 17:07:47.996857 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdmcl\" (UniqueName: \"kubernetes.io/projected/dbfda435-8e90-4baf-ba54-45646e1ca5c9-kube-api-access-xdmcl\") pod \"dbfda435-8e90-4baf-ba54-45646e1ca5c9\" (UID: \"dbfda435-8e90-4baf-ba54-45646e1ca5c9\") " Nov 25 17:07:47 crc kubenswrapper[4802]: I1125 17:07:47.998641 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbfda435-8e90-4baf-ba54-45646e1ca5c9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dbfda435-8e90-4baf-ba54-45646e1ca5c9" (UID: "dbfda435-8e90-4baf-ba54-45646e1ca5c9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:07:48 crc kubenswrapper[4802]: I1125 17:07:48.002511 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbfda435-8e90-4baf-ba54-45646e1ca5c9-kube-api-access-xdmcl" (OuterVolumeSpecName: "kube-api-access-xdmcl") pod "dbfda435-8e90-4baf-ba54-45646e1ca5c9" (UID: "dbfda435-8e90-4baf-ba54-45646e1ca5c9"). InnerVolumeSpecName "kube-api-access-xdmcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:07:48 crc kubenswrapper[4802]: I1125 17:07:48.040116 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-51cf-account-create-update-lhnnv" Nov 25 17:07:48 crc kubenswrapper[4802]: I1125 17:07:48.098426 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e20085a-408e-42db-b120-bfb28eb25b42-operator-scripts\") pod \"2e20085a-408e-42db-b120-bfb28eb25b42\" (UID: \"2e20085a-408e-42db-b120-bfb28eb25b42\") " Nov 25 17:07:48 crc kubenswrapper[4802]: I1125 17:07:48.098470 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxn9f\" (UniqueName: \"kubernetes.io/projected/2e20085a-408e-42db-b120-bfb28eb25b42-kube-api-access-nxn9f\") pod \"2e20085a-408e-42db-b120-bfb28eb25b42\" (UID: \"2e20085a-408e-42db-b120-bfb28eb25b42\") " Nov 25 17:07:48 crc kubenswrapper[4802]: I1125 17:07:48.098760 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dbfda435-8e90-4baf-ba54-45646e1ca5c9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:48 crc kubenswrapper[4802]: I1125 17:07:48.098778 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdmcl\" (UniqueName: \"kubernetes.io/projected/dbfda435-8e90-4baf-ba54-45646e1ca5c9-kube-api-access-xdmcl\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:48 crc kubenswrapper[4802]: I1125 17:07:48.099542 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e20085a-408e-42db-b120-bfb28eb25b42-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2e20085a-408e-42db-b120-bfb28eb25b42" (UID: "2e20085a-408e-42db-b120-bfb28eb25b42"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:07:48 crc kubenswrapper[4802]: I1125 17:07:48.101925 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e20085a-408e-42db-b120-bfb28eb25b42-kube-api-access-nxn9f" (OuterVolumeSpecName: "kube-api-access-nxn9f") pod "2e20085a-408e-42db-b120-bfb28eb25b42" (UID: "2e20085a-408e-42db-b120-bfb28eb25b42"). InnerVolumeSpecName "kube-api-access-nxn9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:07:48 crc kubenswrapper[4802]: I1125 17:07:48.200717 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e20085a-408e-42db-b120-bfb28eb25b42-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:48 crc kubenswrapper[4802]: I1125 17:07:48.201076 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxn9f\" (UniqueName: \"kubernetes.io/projected/2e20085a-408e-42db-b120-bfb28eb25b42-kube-api-access-nxn9f\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:48 crc kubenswrapper[4802]: I1125 17:07:48.644448 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-8sxkq" Nov 25 17:07:48 crc kubenswrapper[4802]: I1125 17:07:48.644436 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-8sxkq" event={"ID":"dbfda435-8e90-4baf-ba54-45646e1ca5c9","Type":"ContainerDied","Data":"acb5ac77ea394d7aa475a8647413e98c2cf693934143fc877c34b5c713f61264"} Nov 25 17:07:48 crc kubenswrapper[4802]: I1125 17:07:48.644505 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="acb5ac77ea394d7aa475a8647413e98c2cf693934143fc877c34b5c713f61264" Nov 25 17:07:48 crc kubenswrapper[4802]: I1125 17:07:48.647034 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-51cf-account-create-update-lhnnv" event={"ID":"2e20085a-408e-42db-b120-bfb28eb25b42","Type":"ContainerDied","Data":"02035075245db14bd28a5710c1fe387e94851d9b3b97d4562711b16f7d2ad92c"} Nov 25 17:07:48 crc kubenswrapper[4802]: I1125 17:07:48.647060 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02035075245db14bd28a5710c1fe387e94851d9b3b97d4562711b16f7d2ad92c" Nov 25 17:07:48 crc kubenswrapper[4802]: I1125 17:07:48.647214 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-51cf-account-create-update-lhnnv" Nov 25 17:07:49 crc kubenswrapper[4802]: I1125 17:07:49.988779 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-86xvv"] Nov 25 17:07:49 crc kubenswrapper[4802]: E1125 17:07:49.989138 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbfda435-8e90-4baf-ba54-45646e1ca5c9" containerName="mariadb-database-create" Nov 25 17:07:49 crc kubenswrapper[4802]: I1125 17:07:49.989152 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbfda435-8e90-4baf-ba54-45646e1ca5c9" containerName="mariadb-database-create" Nov 25 17:07:49 crc kubenswrapper[4802]: E1125 17:07:49.989163 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e20085a-408e-42db-b120-bfb28eb25b42" containerName="mariadb-account-create-update" Nov 25 17:07:49 crc kubenswrapper[4802]: I1125 17:07:49.989172 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e20085a-408e-42db-b120-bfb28eb25b42" containerName="mariadb-account-create-update" Nov 25 17:07:49 crc kubenswrapper[4802]: I1125 17:07:49.989367 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbfda435-8e90-4baf-ba54-45646e1ca5c9" containerName="mariadb-database-create" Nov 25 17:07:49 crc kubenswrapper[4802]: I1125 17:07:49.989383 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e20085a-408e-42db-b120-bfb28eb25b42" containerName="mariadb-account-create-update" Nov 25 17:07:49 crc kubenswrapper[4802]: I1125 17:07:49.989912 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-86xvv" Nov 25 17:07:49 crc kubenswrapper[4802]: I1125 17:07:49.992068 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-5nnrv" Nov 25 17:07:49 crc kubenswrapper[4802]: I1125 17:07:49.992506 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 25 17:07:49 crc kubenswrapper[4802]: I1125 17:07:49.993666 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 25 17:07:50 crc kubenswrapper[4802]: I1125 17:07:50.006283 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-86xvv"] Nov 25 17:07:50 crc kubenswrapper[4802]: I1125 17:07:50.026577 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5fcb\" (UniqueName: \"kubernetes.io/projected/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-kube-api-access-j5fcb\") pod \"glance-db-sync-86xvv\" (UID: \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\") " pod="glance-kuttl-tests/glance-db-sync-86xvv" Nov 25 17:07:50 crc kubenswrapper[4802]: I1125 17:07:50.026697 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-db-sync-config-data\") pod \"glance-db-sync-86xvv\" (UID: \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\") " pod="glance-kuttl-tests/glance-db-sync-86xvv" Nov 25 17:07:50 crc kubenswrapper[4802]: I1125 17:07:50.026741 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-config-data\") pod \"glance-db-sync-86xvv\" (UID: \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\") " pod="glance-kuttl-tests/glance-db-sync-86xvv" Nov 25 17:07:50 crc kubenswrapper[4802]: I1125 17:07:50.026758 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-combined-ca-bundle\") pod \"glance-db-sync-86xvv\" (UID: \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\") " pod="glance-kuttl-tests/glance-db-sync-86xvv" Nov 25 17:07:50 crc kubenswrapper[4802]: I1125 17:07:50.127731 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5fcb\" (UniqueName: \"kubernetes.io/projected/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-kube-api-access-j5fcb\") pod \"glance-db-sync-86xvv\" (UID: \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\") " pod="glance-kuttl-tests/glance-db-sync-86xvv" Nov 25 17:07:50 crc kubenswrapper[4802]: I1125 17:07:50.127866 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-db-sync-config-data\") pod \"glance-db-sync-86xvv\" (UID: \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\") " pod="glance-kuttl-tests/glance-db-sync-86xvv" Nov 25 17:07:50 crc kubenswrapper[4802]: I1125 17:07:50.127925 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-config-data\") pod \"glance-db-sync-86xvv\" (UID: \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\") " pod="glance-kuttl-tests/glance-db-sync-86xvv" Nov 25 17:07:50 crc kubenswrapper[4802]: I1125 17:07:50.127948 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-combined-ca-bundle\") pod \"glance-db-sync-86xvv\" (UID: \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\") " pod="glance-kuttl-tests/glance-db-sync-86xvv" Nov 25 17:07:50 crc kubenswrapper[4802]: I1125 17:07:50.133535 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-combined-ca-bundle\") pod \"glance-db-sync-86xvv\" (UID: \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\") " pod="glance-kuttl-tests/glance-db-sync-86xvv" Nov 25 17:07:50 crc kubenswrapper[4802]: I1125 17:07:50.133565 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-db-sync-config-data\") pod \"glance-db-sync-86xvv\" (UID: \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\") " pod="glance-kuttl-tests/glance-db-sync-86xvv" Nov 25 17:07:50 crc kubenswrapper[4802]: I1125 17:07:50.133592 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-config-data\") pod \"glance-db-sync-86xvv\" (UID: \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\") " pod="glance-kuttl-tests/glance-db-sync-86xvv" Nov 25 17:07:50 crc kubenswrapper[4802]: I1125 17:07:50.147153 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5fcb\" (UniqueName: \"kubernetes.io/projected/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-kube-api-access-j5fcb\") pod \"glance-db-sync-86xvv\" (UID: \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\") " pod="glance-kuttl-tests/glance-db-sync-86xvv" Nov 25 17:07:50 crc kubenswrapper[4802]: I1125 17:07:50.306457 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-86xvv" Nov 25 17:07:50 crc kubenswrapper[4802]: I1125 17:07:50.729837 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-86xvv"] Nov 25 17:07:51 crc kubenswrapper[4802]: I1125 17:07:51.675300 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-86xvv" event={"ID":"1171dd01-c034-46d7-8b54-e80d6c2b7f4c","Type":"ContainerStarted","Data":"02651fd5d18544c34af0b94b5587fac8323de356e54654c775aec55f7ad7e08c"} Nov 25 17:07:51 crc kubenswrapper[4802]: I1125 17:07:51.675668 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-86xvv" event={"ID":"1171dd01-c034-46d7-8b54-e80d6c2b7f4c","Type":"ContainerStarted","Data":"4af323fc612e3f4bbc5970fb1f72af4cbfbf43a1e929eefa3f86c8349798fa3b"} Nov 25 17:07:51 crc kubenswrapper[4802]: I1125 17:07:51.694644 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-86xvv" podStartSLOduration=2.694629043 podStartE2EDuration="2.694629043s" podCreationTimestamp="2025-11-25 17:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:07:51.69116197 +0000 UTC m=+1254.835509156" watchObservedRunningTime="2025-11-25 17:07:51.694629043 +0000 UTC m=+1254.838976229" Nov 25 17:07:54 crc kubenswrapper[4802]: I1125 17:07:54.708929 4802 generic.go:334] "Generic (PLEG): container finished" podID="1171dd01-c034-46d7-8b54-e80d6c2b7f4c" containerID="02651fd5d18544c34af0b94b5587fac8323de356e54654c775aec55f7ad7e08c" exitCode=0 Nov 25 17:07:54 crc kubenswrapper[4802]: I1125 17:07:54.709060 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-86xvv" event={"ID":"1171dd01-c034-46d7-8b54-e80d6c2b7f4c","Type":"ContainerDied","Data":"02651fd5d18544c34af0b94b5587fac8323de356e54654c775aec55f7ad7e08c"} Nov 25 17:07:56 crc kubenswrapper[4802]: I1125 17:07:56.106772 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-86xvv" Nov 25 17:07:56 crc kubenswrapper[4802]: I1125 17:07:56.219796 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-db-sync-config-data\") pod \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\" (UID: \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\") " Nov 25 17:07:56 crc kubenswrapper[4802]: I1125 17:07:56.219922 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-config-data\") pod \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\" (UID: \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\") " Nov 25 17:07:56 crc kubenswrapper[4802]: I1125 17:07:56.219946 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-combined-ca-bundle\") pod \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\" (UID: \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\") " Nov 25 17:07:56 crc kubenswrapper[4802]: I1125 17:07:56.219999 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5fcb\" (UniqueName: \"kubernetes.io/projected/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-kube-api-access-j5fcb\") pod \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\" (UID: \"1171dd01-c034-46d7-8b54-e80d6c2b7f4c\") " Nov 25 17:07:56 crc kubenswrapper[4802]: I1125 17:07:56.227851 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "1171dd01-c034-46d7-8b54-e80d6c2b7f4c" (UID: "1171dd01-c034-46d7-8b54-e80d6c2b7f4c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:07:56 crc kubenswrapper[4802]: I1125 17:07:56.227866 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-kube-api-access-j5fcb" (OuterVolumeSpecName: "kube-api-access-j5fcb") pod "1171dd01-c034-46d7-8b54-e80d6c2b7f4c" (UID: "1171dd01-c034-46d7-8b54-e80d6c2b7f4c"). InnerVolumeSpecName "kube-api-access-j5fcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:07:56 crc kubenswrapper[4802]: I1125 17:07:56.253453 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1171dd01-c034-46d7-8b54-e80d6c2b7f4c" (UID: "1171dd01-c034-46d7-8b54-e80d6c2b7f4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:07:56 crc kubenswrapper[4802]: I1125 17:07:56.278189 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-config-data" (OuterVolumeSpecName: "config-data") pod "1171dd01-c034-46d7-8b54-e80d6c2b7f4c" (UID: "1171dd01-c034-46d7-8b54-e80d6c2b7f4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:07:56 crc kubenswrapper[4802]: I1125 17:07:56.322112 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5fcb\" (UniqueName: \"kubernetes.io/projected/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-kube-api-access-j5fcb\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:56 crc kubenswrapper[4802]: I1125 17:07:56.322180 4802 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:56 crc kubenswrapper[4802]: I1125 17:07:56.322195 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:56 crc kubenswrapper[4802]: I1125 17:07:56.322211 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1171dd01-c034-46d7-8b54-e80d6c2b7f4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 17:07:56 crc kubenswrapper[4802]: I1125 17:07:56.738209 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-86xvv" event={"ID":"1171dd01-c034-46d7-8b54-e80d6c2b7f4c","Type":"ContainerDied","Data":"4af323fc612e3f4bbc5970fb1f72af4cbfbf43a1e929eefa3f86c8349798fa3b"} Nov 25 17:07:56 crc kubenswrapper[4802]: I1125 17:07:56.738262 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-86xvv" Nov 25 17:07:56 crc kubenswrapper[4802]: I1125 17:07:56.738270 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4af323fc612e3f4bbc5970fb1f72af4cbfbf43a1e929eefa3f86c8349798fa3b" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.003834 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 17:07:58 crc kubenswrapper[4802]: E1125 17:07:58.004385 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1171dd01-c034-46d7-8b54-e80d6c2b7f4c" containerName="glance-db-sync" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.004412 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="1171dd01-c034-46d7-8b54-e80d6c2b7f4c" containerName="glance-db-sync" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.004650 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="1171dd01-c034-46d7-8b54-e80d6c2b7f4c" containerName="glance-db-sync" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.005938 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.008282 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.008491 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.012632 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-5nnrv" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.013081 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.013207 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.013320 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.015058 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.051403 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b5c13e1-383a-45b0-baea-4084032c02af-logs\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.051440 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.051471 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.051492 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.051680 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-scripts\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.051719 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b5c13e1-383a-45b0-baea-4084032c02af-httpd-run\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.051814 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.051846 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-config-data\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.051877 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzsmk\" (UniqueName: \"kubernetes.io/projected/2b5c13e1-383a-45b0-baea-4084032c02af-kube-api-access-xzsmk\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.153593 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzsmk\" (UniqueName: \"kubernetes.io/projected/2b5c13e1-383a-45b0-baea-4084032c02af-kube-api-access-xzsmk\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.153852 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b5c13e1-383a-45b0-baea-4084032c02af-logs\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.153874 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.153917 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.153937 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.153966 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-scripts\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.153980 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b5c13e1-383a-45b0-baea-4084032c02af-httpd-run\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.154024 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.154050 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-config-data\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.154558 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b5c13e1-383a-45b0-baea-4084032c02af-logs\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.154630 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b5c13e1-383a-45b0-baea-4084032c02af-httpd-run\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.154814 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.157867 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-scripts\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.158414 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.158666 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.159147 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.159301 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-config-data\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.170933 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzsmk\" (UniqueName: \"kubernetes.io/projected/2b5c13e1-383a-45b0-baea-4084032c02af-kube-api-access-xzsmk\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.175147 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:58 crc kubenswrapper[4802]: I1125 17:07:58.324739 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:07:59 crc kubenswrapper[4802]: I1125 17:07:59.463367 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 17:07:59 crc kubenswrapper[4802]: I1125 17:07:59.766383 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2b5c13e1-383a-45b0-baea-4084032c02af","Type":"ContainerStarted","Data":"961245ee8c600e26280e0e048216bc66f3c52b80102484ef19441164189c31d8"} Nov 25 17:08:00 crc kubenswrapper[4802]: I1125 17:08:00.778034 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2b5c13e1-383a-45b0-baea-4084032c02af","Type":"ContainerStarted","Data":"9cc305a24e0b2a62c34ab557d50c400f9979b4c9b3625e90cd860c58563a13a7"} Nov 25 17:08:00 crc kubenswrapper[4802]: I1125 17:08:00.778434 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2b5c13e1-383a-45b0-baea-4084032c02af","Type":"ContainerStarted","Data":"241d7bafb8dc959876032bbfa43c991bd586d3b1961ff433aa32d4e85a5237d2"} Nov 25 17:08:00 crc kubenswrapper[4802]: I1125 17:08:00.815355 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.815326797 podStartE2EDuration="3.815326797s" podCreationTimestamp="2025-11-25 17:07:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:08:00.805205095 +0000 UTC m=+1263.949552321" watchObservedRunningTime="2025-11-25 17:08:00.815326797 +0000 UTC m=+1263.959674023" Nov 25 17:08:08 crc kubenswrapper[4802]: I1125 17:08:08.325614 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:08:08 crc kubenswrapper[4802]: I1125 17:08:08.326171 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:08:08 crc kubenswrapper[4802]: I1125 17:08:08.356921 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:08:08 crc kubenswrapper[4802]: I1125 17:08:08.379173 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:08:08 crc kubenswrapper[4802]: I1125 17:08:08.843315 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:08:08 crc kubenswrapper[4802]: I1125 17:08:08.843362 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:08:10 crc kubenswrapper[4802]: I1125 17:08:10.817202 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:08:10 crc kubenswrapper[4802]: I1125 17:08:10.825645 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:08:12 crc kubenswrapper[4802]: I1125 17:08:12.199510 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-86xvv"] Nov 25 17:08:12 crc kubenswrapper[4802]: I1125 17:08:12.206028 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-86xvv"] Nov 25 17:08:12 crc kubenswrapper[4802]: I1125 17:08:12.251468 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance51cf-account-delete-dr4xz"] Nov 25 17:08:12 crc kubenswrapper[4802]: I1125 17:08:12.252502 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance51cf-account-delete-dr4xz" Nov 25 17:08:12 crc kubenswrapper[4802]: I1125 17:08:12.264618 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance51cf-account-delete-dr4xz"] Nov 25 17:08:12 crc kubenswrapper[4802]: I1125 17:08:12.302022 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 17:08:12 crc kubenswrapper[4802]: I1125 17:08:12.373679 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpb2t\" (UniqueName: \"kubernetes.io/projected/47d2e8a6-007a-4c2c-8539-f66d016abba9-kube-api-access-xpb2t\") pod \"glance51cf-account-delete-dr4xz\" (UID: \"47d2e8a6-007a-4c2c-8539-f66d016abba9\") " pod="glance-kuttl-tests/glance51cf-account-delete-dr4xz" Nov 25 17:08:12 crc kubenswrapper[4802]: I1125 17:08:12.373722 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47d2e8a6-007a-4c2c-8539-f66d016abba9-operator-scripts\") pod \"glance51cf-account-delete-dr4xz\" (UID: \"47d2e8a6-007a-4c2c-8539-f66d016abba9\") " pod="glance-kuttl-tests/glance51cf-account-delete-dr4xz" Nov 25 17:08:12 crc kubenswrapper[4802]: I1125 17:08:12.475685 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpb2t\" (UniqueName: \"kubernetes.io/projected/47d2e8a6-007a-4c2c-8539-f66d016abba9-kube-api-access-xpb2t\") pod \"glance51cf-account-delete-dr4xz\" (UID: \"47d2e8a6-007a-4c2c-8539-f66d016abba9\") " pod="glance-kuttl-tests/glance51cf-account-delete-dr4xz" Nov 25 17:08:12 crc kubenswrapper[4802]: I1125 17:08:12.475733 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47d2e8a6-007a-4c2c-8539-f66d016abba9-operator-scripts\") pod \"glance51cf-account-delete-dr4xz\" (UID: \"47d2e8a6-007a-4c2c-8539-f66d016abba9\") " pod="glance-kuttl-tests/glance51cf-account-delete-dr4xz" Nov 25 17:08:12 crc kubenswrapper[4802]: I1125 17:08:12.476516 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47d2e8a6-007a-4c2c-8539-f66d016abba9-operator-scripts\") pod \"glance51cf-account-delete-dr4xz\" (UID: \"47d2e8a6-007a-4c2c-8539-f66d016abba9\") " pod="glance-kuttl-tests/glance51cf-account-delete-dr4xz" Nov 25 17:08:12 crc kubenswrapper[4802]: I1125 17:08:12.502181 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpb2t\" (UniqueName: \"kubernetes.io/projected/47d2e8a6-007a-4c2c-8539-f66d016abba9-kube-api-access-xpb2t\") pod \"glance51cf-account-delete-dr4xz\" (UID: \"47d2e8a6-007a-4c2c-8539-f66d016abba9\") " pod="glance-kuttl-tests/glance51cf-account-delete-dr4xz" Nov 25 17:08:12 crc kubenswrapper[4802]: I1125 17:08:12.573037 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance51cf-account-delete-dr4xz" Nov 25 17:08:12 crc kubenswrapper[4802]: I1125 17:08:12.872074 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="2b5c13e1-383a-45b0-baea-4084032c02af" containerName="glance-log" containerID="cri-o://241d7bafb8dc959876032bbfa43c991bd586d3b1961ff433aa32d4e85a5237d2" gracePeriod=30 Nov 25 17:08:12 crc kubenswrapper[4802]: I1125 17:08:12.872219 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="2b5c13e1-383a-45b0-baea-4084032c02af" containerName="glance-httpd" containerID="cri-o://9cc305a24e0b2a62c34ab557d50c400f9979b4c9b3625e90cd860c58563a13a7" gracePeriod=30 Nov 25 17:08:12 crc kubenswrapper[4802]: I1125 17:08:12.877764 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="2b5c13e1-383a-45b0-baea-4084032c02af" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.110:9292/healthcheck\": EOF" Nov 25 17:08:13 crc kubenswrapper[4802]: I1125 17:08:13.002480 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance51cf-account-delete-dr4xz"] Nov 25 17:08:13 crc kubenswrapper[4802]: W1125 17:08:13.004599 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47d2e8a6_007a_4c2c_8539_f66d016abba9.slice/crio-e2094caaddf2f488825de6202c29fb6a30758d7ff014d1d482bcf97ff5337857 WatchSource:0}: Error finding container e2094caaddf2f488825de6202c29fb6a30758d7ff014d1d482bcf97ff5337857: Status 404 returned error can't find the container with id e2094caaddf2f488825de6202c29fb6a30758d7ff014d1d482bcf97ff5337857 Nov 25 17:08:13 crc kubenswrapper[4802]: I1125 17:08:13.513315 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1171dd01-c034-46d7-8b54-e80d6c2b7f4c" path="/var/lib/kubelet/pods/1171dd01-c034-46d7-8b54-e80d6c2b7f4c/volumes" Nov 25 17:08:13 crc kubenswrapper[4802]: I1125 17:08:13.883363 4802 generic.go:334] "Generic (PLEG): container finished" podID="47d2e8a6-007a-4c2c-8539-f66d016abba9" containerID="5c0e9c2c113faea7b3e3112544b28b3ea5a43e4b6e94989a8c183a64a79b22d1" exitCode=0 Nov 25 17:08:13 crc kubenswrapper[4802]: I1125 17:08:13.883455 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance51cf-account-delete-dr4xz" event={"ID":"47d2e8a6-007a-4c2c-8539-f66d016abba9","Type":"ContainerDied","Data":"5c0e9c2c113faea7b3e3112544b28b3ea5a43e4b6e94989a8c183a64a79b22d1"} Nov 25 17:08:13 crc kubenswrapper[4802]: I1125 17:08:13.883543 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance51cf-account-delete-dr4xz" event={"ID":"47d2e8a6-007a-4c2c-8539-f66d016abba9","Type":"ContainerStarted","Data":"e2094caaddf2f488825de6202c29fb6a30758d7ff014d1d482bcf97ff5337857"} Nov 25 17:08:13 crc kubenswrapper[4802]: I1125 17:08:13.887294 4802 generic.go:334] "Generic (PLEG): container finished" podID="2b5c13e1-383a-45b0-baea-4084032c02af" containerID="241d7bafb8dc959876032bbfa43c991bd586d3b1961ff433aa32d4e85a5237d2" exitCode=143 Nov 25 17:08:13 crc kubenswrapper[4802]: I1125 17:08:13.887363 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2b5c13e1-383a-45b0-baea-4084032c02af","Type":"ContainerDied","Data":"241d7bafb8dc959876032bbfa43c991bd586d3b1961ff433aa32d4e85a5237d2"} Nov 25 17:08:15 crc kubenswrapper[4802]: I1125 17:08:15.197821 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance51cf-account-delete-dr4xz" Nov 25 17:08:15 crc kubenswrapper[4802]: I1125 17:08:15.326981 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpb2t\" (UniqueName: \"kubernetes.io/projected/47d2e8a6-007a-4c2c-8539-f66d016abba9-kube-api-access-xpb2t\") pod \"47d2e8a6-007a-4c2c-8539-f66d016abba9\" (UID: \"47d2e8a6-007a-4c2c-8539-f66d016abba9\") " Nov 25 17:08:15 crc kubenswrapper[4802]: I1125 17:08:15.327212 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47d2e8a6-007a-4c2c-8539-f66d016abba9-operator-scripts\") pod \"47d2e8a6-007a-4c2c-8539-f66d016abba9\" (UID: \"47d2e8a6-007a-4c2c-8539-f66d016abba9\") " Nov 25 17:08:15 crc kubenswrapper[4802]: I1125 17:08:15.328107 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47d2e8a6-007a-4c2c-8539-f66d016abba9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "47d2e8a6-007a-4c2c-8539-f66d016abba9" (UID: "47d2e8a6-007a-4c2c-8539-f66d016abba9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:08:15 crc kubenswrapper[4802]: I1125 17:08:15.332602 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47d2e8a6-007a-4c2c-8539-f66d016abba9-kube-api-access-xpb2t" (OuterVolumeSpecName: "kube-api-access-xpb2t") pod "47d2e8a6-007a-4c2c-8539-f66d016abba9" (UID: "47d2e8a6-007a-4c2c-8539-f66d016abba9"). InnerVolumeSpecName "kube-api-access-xpb2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:08:15 crc kubenswrapper[4802]: I1125 17:08:15.429429 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpb2t\" (UniqueName: \"kubernetes.io/projected/47d2e8a6-007a-4c2c-8539-f66d016abba9-kube-api-access-xpb2t\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:15 crc kubenswrapper[4802]: I1125 17:08:15.429476 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47d2e8a6-007a-4c2c-8539-f66d016abba9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:15 crc kubenswrapper[4802]: I1125 17:08:15.907204 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance51cf-account-delete-dr4xz" event={"ID":"47d2e8a6-007a-4c2c-8539-f66d016abba9","Type":"ContainerDied","Data":"e2094caaddf2f488825de6202c29fb6a30758d7ff014d1d482bcf97ff5337857"} Nov 25 17:08:15 crc kubenswrapper[4802]: I1125 17:08:15.907246 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2094caaddf2f488825de6202c29fb6a30758d7ff014d1d482bcf97ff5337857" Nov 25 17:08:15 crc kubenswrapper[4802]: I1125 17:08:15.907301 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance51cf-account-delete-dr4xz" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.388871 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.543956 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-scripts\") pod \"2b5c13e1-383a-45b0-baea-4084032c02af\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.544314 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzsmk\" (UniqueName: \"kubernetes.io/projected/2b5c13e1-383a-45b0-baea-4084032c02af-kube-api-access-xzsmk\") pod \"2b5c13e1-383a-45b0-baea-4084032c02af\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.544341 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b5c13e1-383a-45b0-baea-4084032c02af-httpd-run\") pod \"2b5c13e1-383a-45b0-baea-4084032c02af\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.544852 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b5c13e1-383a-45b0-baea-4084032c02af-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2b5c13e1-383a-45b0-baea-4084032c02af" (UID: "2b5c13e1-383a-45b0-baea-4084032c02af"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.544930 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"2b5c13e1-383a-45b0-baea-4084032c02af\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.544971 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-config-data\") pod \"2b5c13e1-383a-45b0-baea-4084032c02af\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.544990 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-internal-tls-certs\") pod \"2b5c13e1-383a-45b0-baea-4084032c02af\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.545022 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b5c13e1-383a-45b0-baea-4084032c02af-logs\") pod \"2b5c13e1-383a-45b0-baea-4084032c02af\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.545037 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-combined-ca-bundle\") pod \"2b5c13e1-383a-45b0-baea-4084032c02af\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.545092 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-public-tls-certs\") pod \"2b5c13e1-383a-45b0-baea-4084032c02af\" (UID: \"2b5c13e1-383a-45b0-baea-4084032c02af\") " Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.545375 4802 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b5c13e1-383a-45b0-baea-4084032c02af-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.545836 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b5c13e1-383a-45b0-baea-4084032c02af-logs" (OuterVolumeSpecName: "logs") pod "2b5c13e1-383a-45b0-baea-4084032c02af" (UID: "2b5c13e1-383a-45b0-baea-4084032c02af"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.550168 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "2b5c13e1-383a-45b0-baea-4084032c02af" (UID: "2b5c13e1-383a-45b0-baea-4084032c02af"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.550176 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b5c13e1-383a-45b0-baea-4084032c02af-kube-api-access-xzsmk" (OuterVolumeSpecName: "kube-api-access-xzsmk") pod "2b5c13e1-383a-45b0-baea-4084032c02af" (UID: "2b5c13e1-383a-45b0-baea-4084032c02af"). InnerVolumeSpecName "kube-api-access-xzsmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.550523 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-scripts" (OuterVolumeSpecName: "scripts") pod "2b5c13e1-383a-45b0-baea-4084032c02af" (UID: "2b5c13e1-383a-45b0-baea-4084032c02af"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.564991 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2b5c13e1-383a-45b0-baea-4084032c02af" (UID: "2b5c13e1-383a-45b0-baea-4084032c02af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.579560 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-config-data" (OuterVolumeSpecName: "config-data") pod "2b5c13e1-383a-45b0-baea-4084032c02af" (UID: "2b5c13e1-383a-45b0-baea-4084032c02af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.580986 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2b5c13e1-383a-45b0-baea-4084032c02af" (UID: "2b5c13e1-383a-45b0-baea-4084032c02af"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.589685 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2b5c13e1-383a-45b0-baea-4084032c02af" (UID: "2b5c13e1-383a-45b0-baea-4084032c02af"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.646256 4802 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.646299 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.646314 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzsmk\" (UniqueName: \"kubernetes.io/projected/2b5c13e1-383a-45b0-baea-4084032c02af-kube-api-access-xzsmk\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.646354 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.646367 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.646378 4802 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.646389 4802 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b5c13e1-383a-45b0-baea-4084032c02af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.646401 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b5c13e1-383a-45b0-baea-4084032c02af-logs\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.658928 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.748222 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.917166 4802 generic.go:334] "Generic (PLEG): container finished" podID="2b5c13e1-383a-45b0-baea-4084032c02af" containerID="9cc305a24e0b2a62c34ab557d50c400f9979b4c9b3625e90cd860c58563a13a7" exitCode=0 Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.917210 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2b5c13e1-383a-45b0-baea-4084032c02af","Type":"ContainerDied","Data":"9cc305a24e0b2a62c34ab557d50c400f9979b4c9b3625e90cd860c58563a13a7"} Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.917239 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2b5c13e1-383a-45b0-baea-4084032c02af","Type":"ContainerDied","Data":"961245ee8c600e26280e0e048216bc66f3c52b80102484ef19441164189c31d8"} Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.917255 4802 scope.go:117] "RemoveContainer" containerID="9cc305a24e0b2a62c34ab557d50c400f9979b4c9b3625e90cd860c58563a13a7" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.917824 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.937480 4802 scope.go:117] "RemoveContainer" containerID="241d7bafb8dc959876032bbfa43c991bd586d3b1961ff433aa32d4e85a5237d2" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.955800 4802 scope.go:117] "RemoveContainer" containerID="9cc305a24e0b2a62c34ab557d50c400f9979b4c9b3625e90cd860c58563a13a7" Nov 25 17:08:16 crc kubenswrapper[4802]: E1125 17:08:16.956571 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cc305a24e0b2a62c34ab557d50c400f9979b4c9b3625e90cd860c58563a13a7\": container with ID starting with 9cc305a24e0b2a62c34ab557d50c400f9979b4c9b3625e90cd860c58563a13a7 not found: ID does not exist" containerID="9cc305a24e0b2a62c34ab557d50c400f9979b4c9b3625e90cd860c58563a13a7" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.956622 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cc305a24e0b2a62c34ab557d50c400f9979b4c9b3625e90cd860c58563a13a7"} err="failed to get container status \"9cc305a24e0b2a62c34ab557d50c400f9979b4c9b3625e90cd860c58563a13a7\": rpc error: code = NotFound desc = could not find container \"9cc305a24e0b2a62c34ab557d50c400f9979b4c9b3625e90cd860c58563a13a7\": container with ID starting with 9cc305a24e0b2a62c34ab557d50c400f9979b4c9b3625e90cd860c58563a13a7 not found: ID does not exist" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.956659 4802 scope.go:117] "RemoveContainer" containerID="241d7bafb8dc959876032bbfa43c991bd586d3b1961ff433aa32d4e85a5237d2" Nov 25 17:08:16 crc kubenswrapper[4802]: E1125 17:08:16.957165 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"241d7bafb8dc959876032bbfa43c991bd586d3b1961ff433aa32d4e85a5237d2\": container with ID starting with 241d7bafb8dc959876032bbfa43c991bd586d3b1961ff433aa32d4e85a5237d2 not found: ID does not exist" containerID="241d7bafb8dc959876032bbfa43c991bd586d3b1961ff433aa32d4e85a5237d2" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.957206 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"241d7bafb8dc959876032bbfa43c991bd586d3b1961ff433aa32d4e85a5237d2"} err="failed to get container status \"241d7bafb8dc959876032bbfa43c991bd586d3b1961ff433aa32d4e85a5237d2\": rpc error: code = NotFound desc = could not find container \"241d7bafb8dc959876032bbfa43c991bd586d3b1961ff433aa32d4e85a5237d2\": container with ID starting with 241d7bafb8dc959876032bbfa43c991bd586d3b1961ff433aa32d4e85a5237d2 not found: ID does not exist" Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.962765 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 17:08:16 crc kubenswrapper[4802]: I1125 17:08:16.969719 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.273500 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-8sxkq"] Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.299352 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-8sxkq"] Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.305933 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance51cf-account-delete-dr4xz"] Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.312755 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-51cf-account-create-update-lhnnv"] Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.320275 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-51cf-account-create-update-lhnnv"] Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.325670 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance51cf-account-delete-dr4xz"] Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.512280 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b5c13e1-383a-45b0-baea-4084032c02af" path="/var/lib/kubelet/pods/2b5c13e1-383a-45b0-baea-4084032c02af/volumes" Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.512990 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e20085a-408e-42db-b120-bfb28eb25b42" path="/var/lib/kubelet/pods/2e20085a-408e-42db-b120-bfb28eb25b42/volumes" Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.513512 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47d2e8a6-007a-4c2c-8539-f66d016abba9" path="/var/lib/kubelet/pods/47d2e8a6-007a-4c2c-8539-f66d016abba9/volumes" Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.514491 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbfda435-8e90-4baf-ba54-45646e1ca5c9" path="/var/lib/kubelet/pods/dbfda435-8e90-4baf-ba54-45646e1ca5c9/volumes" Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.966276 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-qbqgx"] Nov 25 17:08:17 crc kubenswrapper[4802]: E1125 17:08:17.966713 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47d2e8a6-007a-4c2c-8539-f66d016abba9" containerName="mariadb-account-delete" Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.966744 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="47d2e8a6-007a-4c2c-8539-f66d016abba9" containerName="mariadb-account-delete" Nov 25 17:08:17 crc kubenswrapper[4802]: E1125 17:08:17.966774 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b5c13e1-383a-45b0-baea-4084032c02af" containerName="glance-log" Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.966786 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b5c13e1-383a-45b0-baea-4084032c02af" containerName="glance-log" Nov 25 17:08:17 crc kubenswrapper[4802]: E1125 17:08:17.966814 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b5c13e1-383a-45b0-baea-4084032c02af" containerName="glance-httpd" Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.966826 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b5c13e1-383a-45b0-baea-4084032c02af" containerName="glance-httpd" Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.967050 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b5c13e1-383a-45b0-baea-4084032c02af" containerName="glance-httpd" Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.967083 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="47d2e8a6-007a-4c2c-8539-f66d016abba9" containerName="mariadb-account-delete" Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.967141 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b5c13e1-383a-45b0-baea-4084032c02af" containerName="glance-log" Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.968110 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-qbqgx" Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.974836 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-qbqgx"] Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.979909 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr"] Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.980822 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr" Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.982728 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 25 17:08:17 crc kubenswrapper[4802]: I1125 17:08:17.986904 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr"] Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.166553 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e34282e2-a266-4d67-ae13-0bce1083ae91-operator-scripts\") pod \"glance-d3b4-account-create-update-h4wcr\" (UID: \"e34282e2-a266-4d67-ae13-0bce1083ae91\") " pod="glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr" Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.166652 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a38f8ea1-50bb-4b06-883c-9be588679c41-operator-scripts\") pod \"glance-db-create-qbqgx\" (UID: \"a38f8ea1-50bb-4b06-883c-9be588679c41\") " pod="glance-kuttl-tests/glance-db-create-qbqgx" Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.166949 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n7mc\" (UniqueName: \"kubernetes.io/projected/a38f8ea1-50bb-4b06-883c-9be588679c41-kube-api-access-6n7mc\") pod \"glance-db-create-qbqgx\" (UID: \"a38f8ea1-50bb-4b06-883c-9be588679c41\") " pod="glance-kuttl-tests/glance-db-create-qbqgx" Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.167053 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jkhj\" (UniqueName: \"kubernetes.io/projected/e34282e2-a266-4d67-ae13-0bce1083ae91-kube-api-access-2jkhj\") pod \"glance-d3b4-account-create-update-h4wcr\" (UID: \"e34282e2-a266-4d67-ae13-0bce1083ae91\") " pod="glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr" Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.268576 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jkhj\" (UniqueName: \"kubernetes.io/projected/e34282e2-a266-4d67-ae13-0bce1083ae91-kube-api-access-2jkhj\") pod \"glance-d3b4-account-create-update-h4wcr\" (UID: \"e34282e2-a266-4d67-ae13-0bce1083ae91\") " pod="glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr" Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.268648 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e34282e2-a266-4d67-ae13-0bce1083ae91-operator-scripts\") pod \"glance-d3b4-account-create-update-h4wcr\" (UID: \"e34282e2-a266-4d67-ae13-0bce1083ae91\") " pod="glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr" Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.268672 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a38f8ea1-50bb-4b06-883c-9be588679c41-operator-scripts\") pod \"glance-db-create-qbqgx\" (UID: \"a38f8ea1-50bb-4b06-883c-9be588679c41\") " pod="glance-kuttl-tests/glance-db-create-qbqgx" Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.268711 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n7mc\" (UniqueName: \"kubernetes.io/projected/a38f8ea1-50bb-4b06-883c-9be588679c41-kube-api-access-6n7mc\") pod \"glance-db-create-qbqgx\" (UID: \"a38f8ea1-50bb-4b06-883c-9be588679c41\") " pod="glance-kuttl-tests/glance-db-create-qbqgx" Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.269787 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a38f8ea1-50bb-4b06-883c-9be588679c41-operator-scripts\") pod \"glance-db-create-qbqgx\" (UID: \"a38f8ea1-50bb-4b06-883c-9be588679c41\") " pod="glance-kuttl-tests/glance-db-create-qbqgx" Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.270164 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e34282e2-a266-4d67-ae13-0bce1083ae91-operator-scripts\") pod \"glance-d3b4-account-create-update-h4wcr\" (UID: \"e34282e2-a266-4d67-ae13-0bce1083ae91\") " pod="glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr" Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.285691 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jkhj\" (UniqueName: \"kubernetes.io/projected/e34282e2-a266-4d67-ae13-0bce1083ae91-kube-api-access-2jkhj\") pod \"glance-d3b4-account-create-update-h4wcr\" (UID: \"e34282e2-a266-4d67-ae13-0bce1083ae91\") " pod="glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr" Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.290543 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n7mc\" (UniqueName: \"kubernetes.io/projected/a38f8ea1-50bb-4b06-883c-9be588679c41-kube-api-access-6n7mc\") pod \"glance-db-create-qbqgx\" (UID: \"a38f8ea1-50bb-4b06-883c-9be588679c41\") " pod="glance-kuttl-tests/glance-db-create-qbqgx" Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.295169 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-qbqgx" Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.311935 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr" Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.718024 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-qbqgx"] Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.773736 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr"] Nov 25 17:08:18 crc kubenswrapper[4802]: W1125 17:08:18.777811 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode34282e2_a266_4d67_ae13_0bce1083ae91.slice/crio-04f14327f2a26954bcd60969eee201c7769f686d04878f04d42d7a856db58fc0 WatchSource:0}: Error finding container 04f14327f2a26954bcd60969eee201c7769f686d04878f04d42d7a856db58fc0: Status 404 returned error can't find the container with id 04f14327f2a26954bcd60969eee201c7769f686d04878f04d42d7a856db58fc0 Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.934940 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr" event={"ID":"e34282e2-a266-4d67-ae13-0bce1083ae91","Type":"ContainerStarted","Data":"0507e27121ac8217fc618d1b0fd8237f3553d09abcdc07eb1cb7e543c8f29df4"} Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.935320 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr" event={"ID":"e34282e2-a266-4d67-ae13-0bce1083ae91","Type":"ContainerStarted","Data":"04f14327f2a26954bcd60969eee201c7769f686d04878f04d42d7a856db58fc0"} Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.937831 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-qbqgx" event={"ID":"a38f8ea1-50bb-4b06-883c-9be588679c41","Type":"ContainerStarted","Data":"ce9957fb5a14f6a8b87b508a6be9ee212b4a35a6d38b842c0f8cc619c05341e4"} Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.937881 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-qbqgx" event={"ID":"a38f8ea1-50bb-4b06-883c-9be588679c41","Type":"ContainerStarted","Data":"d91edf91da027bc68f58c75fc5f0c3e6bdb736f04cd3a78cf90c4dcb25a5d1c4"} Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.955475 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr" podStartSLOduration=1.9554526779999999 podStartE2EDuration="1.955452678s" podCreationTimestamp="2025-11-25 17:08:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:08:18.949518778 +0000 UTC m=+1282.093865994" watchObservedRunningTime="2025-11-25 17:08:18.955452678 +0000 UTC m=+1282.099799874" Nov 25 17:08:18 crc kubenswrapper[4802]: I1125 17:08:18.968649 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-create-qbqgx" podStartSLOduration=1.968626683 podStartE2EDuration="1.968626683s" podCreationTimestamp="2025-11-25 17:08:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:08:18.962438486 +0000 UTC m=+1282.106785682" watchObservedRunningTime="2025-11-25 17:08:18.968626683 +0000 UTC m=+1282.112973889" Nov 25 17:08:20 crc kubenswrapper[4802]: I1125 17:08:20.904790 4802 generic.go:334] "Generic (PLEG): container finished" podID="e34282e2-a266-4d67-ae13-0bce1083ae91" containerID="0507e27121ac8217fc618d1b0fd8237f3553d09abcdc07eb1cb7e543c8f29df4" exitCode=0 Nov 25 17:08:20 crc kubenswrapper[4802]: I1125 17:08:20.904891 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr" event={"ID":"e34282e2-a266-4d67-ae13-0bce1083ae91","Type":"ContainerDied","Data":"0507e27121ac8217fc618d1b0fd8237f3553d09abcdc07eb1cb7e543c8f29df4"} Nov 25 17:08:20 crc kubenswrapper[4802]: I1125 17:08:20.909458 4802 generic.go:334] "Generic (PLEG): container finished" podID="a38f8ea1-50bb-4b06-883c-9be588679c41" containerID="ce9957fb5a14f6a8b87b508a6be9ee212b4a35a6d38b842c0f8cc619c05341e4" exitCode=0 Nov 25 17:08:20 crc kubenswrapper[4802]: I1125 17:08:20.909629 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-qbqgx" event={"ID":"a38f8ea1-50bb-4b06-883c-9be588679c41","Type":"ContainerDied","Data":"ce9957fb5a14f6a8b87b508a6be9ee212b4a35a6d38b842c0f8cc619c05341e4"} Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.302542 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr" Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.310356 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-qbqgx" Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.377156 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n7mc\" (UniqueName: \"kubernetes.io/projected/a38f8ea1-50bb-4b06-883c-9be588679c41-kube-api-access-6n7mc\") pod \"a38f8ea1-50bb-4b06-883c-9be588679c41\" (UID: \"a38f8ea1-50bb-4b06-883c-9be588679c41\") " Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.377253 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jkhj\" (UniqueName: \"kubernetes.io/projected/e34282e2-a266-4d67-ae13-0bce1083ae91-kube-api-access-2jkhj\") pod \"e34282e2-a266-4d67-ae13-0bce1083ae91\" (UID: \"e34282e2-a266-4d67-ae13-0bce1083ae91\") " Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.377296 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a38f8ea1-50bb-4b06-883c-9be588679c41-operator-scripts\") pod \"a38f8ea1-50bb-4b06-883c-9be588679c41\" (UID: \"a38f8ea1-50bb-4b06-883c-9be588679c41\") " Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.377421 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e34282e2-a266-4d67-ae13-0bce1083ae91-operator-scripts\") pod \"e34282e2-a266-4d67-ae13-0bce1083ae91\" (UID: \"e34282e2-a266-4d67-ae13-0bce1083ae91\") " Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.378170 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e34282e2-a266-4d67-ae13-0bce1083ae91-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e34282e2-a266-4d67-ae13-0bce1083ae91" (UID: "e34282e2-a266-4d67-ae13-0bce1083ae91"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.378660 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a38f8ea1-50bb-4b06-883c-9be588679c41-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a38f8ea1-50bb-4b06-883c-9be588679c41" (UID: "a38f8ea1-50bb-4b06-883c-9be588679c41"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.383105 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a38f8ea1-50bb-4b06-883c-9be588679c41-kube-api-access-6n7mc" (OuterVolumeSpecName: "kube-api-access-6n7mc") pod "a38f8ea1-50bb-4b06-883c-9be588679c41" (UID: "a38f8ea1-50bb-4b06-883c-9be588679c41"). InnerVolumeSpecName "kube-api-access-6n7mc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.383175 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e34282e2-a266-4d67-ae13-0bce1083ae91-kube-api-access-2jkhj" (OuterVolumeSpecName: "kube-api-access-2jkhj") pod "e34282e2-a266-4d67-ae13-0bce1083ae91" (UID: "e34282e2-a266-4d67-ae13-0bce1083ae91"). InnerVolumeSpecName "kube-api-access-2jkhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.479157 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jkhj\" (UniqueName: \"kubernetes.io/projected/e34282e2-a266-4d67-ae13-0bce1083ae91-kube-api-access-2jkhj\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.479185 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a38f8ea1-50bb-4b06-883c-9be588679c41-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.479195 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e34282e2-a266-4d67-ae13-0bce1083ae91-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.479204 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n7mc\" (UniqueName: \"kubernetes.io/projected/a38f8ea1-50bb-4b06-883c-9be588679c41-kube-api-access-6n7mc\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.935372 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-qbqgx" event={"ID":"a38f8ea1-50bb-4b06-883c-9be588679c41","Type":"ContainerDied","Data":"d91edf91da027bc68f58c75fc5f0c3e6bdb736f04cd3a78cf90c4dcb25a5d1c4"} Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.935426 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d91edf91da027bc68f58c75fc5f0c3e6bdb736f04cd3a78cf90c4dcb25a5d1c4" Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.935469 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-qbqgx" Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.937839 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr" event={"ID":"e34282e2-a266-4d67-ae13-0bce1083ae91","Type":"ContainerDied","Data":"04f14327f2a26954bcd60969eee201c7769f686d04878f04d42d7a856db58fc0"} Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.937868 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04f14327f2a26954bcd60969eee201c7769f686d04878f04d42d7a856db58fc0" Nov 25 17:08:22 crc kubenswrapper[4802]: I1125 17:08:22.937928 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.204735 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-wc8x4"] Nov 25 17:08:23 crc kubenswrapper[4802]: E1125 17:08:23.205066 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a38f8ea1-50bb-4b06-883c-9be588679c41" containerName="mariadb-database-create" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.205080 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a38f8ea1-50bb-4b06-883c-9be588679c41" containerName="mariadb-database-create" Nov 25 17:08:23 crc kubenswrapper[4802]: E1125 17:08:23.205107 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e34282e2-a266-4d67-ae13-0bce1083ae91" containerName="mariadb-account-create-update" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.205115 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="e34282e2-a266-4d67-ae13-0bce1083ae91" containerName="mariadb-account-create-update" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.205293 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a38f8ea1-50bb-4b06-883c-9be588679c41" containerName="mariadb-database-create" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.205314 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="e34282e2-a266-4d67-ae13-0bce1083ae91" containerName="mariadb-account-create-update" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.205825 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-wc8x4" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.208260 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.208463 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-87pxg" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.216107 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-wc8x4"] Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.292481 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dde71ad-bcda-407e-a490-e8323685e7eb-config-data\") pod \"glance-db-sync-wc8x4\" (UID: \"2dde71ad-bcda-407e-a490-e8323685e7eb\") " pod="glance-kuttl-tests/glance-db-sync-wc8x4" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.292546 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2dde71ad-bcda-407e-a490-e8323685e7eb-db-sync-config-data\") pod \"glance-db-sync-wc8x4\" (UID: \"2dde71ad-bcda-407e-a490-e8323685e7eb\") " pod="glance-kuttl-tests/glance-db-sync-wc8x4" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.292632 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ln87\" (UniqueName: \"kubernetes.io/projected/2dde71ad-bcda-407e-a490-e8323685e7eb-kube-api-access-6ln87\") pod \"glance-db-sync-wc8x4\" (UID: \"2dde71ad-bcda-407e-a490-e8323685e7eb\") " pod="glance-kuttl-tests/glance-db-sync-wc8x4" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.394101 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dde71ad-bcda-407e-a490-e8323685e7eb-config-data\") pod \"glance-db-sync-wc8x4\" (UID: \"2dde71ad-bcda-407e-a490-e8323685e7eb\") " pod="glance-kuttl-tests/glance-db-sync-wc8x4" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.394195 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2dde71ad-bcda-407e-a490-e8323685e7eb-db-sync-config-data\") pod \"glance-db-sync-wc8x4\" (UID: \"2dde71ad-bcda-407e-a490-e8323685e7eb\") " pod="glance-kuttl-tests/glance-db-sync-wc8x4" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.394243 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ln87\" (UniqueName: \"kubernetes.io/projected/2dde71ad-bcda-407e-a490-e8323685e7eb-kube-api-access-6ln87\") pod \"glance-db-sync-wc8x4\" (UID: \"2dde71ad-bcda-407e-a490-e8323685e7eb\") " pod="glance-kuttl-tests/glance-db-sync-wc8x4" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.406881 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dde71ad-bcda-407e-a490-e8323685e7eb-config-data\") pod \"glance-db-sync-wc8x4\" (UID: \"2dde71ad-bcda-407e-a490-e8323685e7eb\") " pod="glance-kuttl-tests/glance-db-sync-wc8x4" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.406869 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2dde71ad-bcda-407e-a490-e8323685e7eb-db-sync-config-data\") pod \"glance-db-sync-wc8x4\" (UID: \"2dde71ad-bcda-407e-a490-e8323685e7eb\") " pod="glance-kuttl-tests/glance-db-sync-wc8x4" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.415588 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ln87\" (UniqueName: \"kubernetes.io/projected/2dde71ad-bcda-407e-a490-e8323685e7eb-kube-api-access-6ln87\") pod \"glance-db-sync-wc8x4\" (UID: \"2dde71ad-bcda-407e-a490-e8323685e7eb\") " pod="glance-kuttl-tests/glance-db-sync-wc8x4" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.553194 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-wc8x4" Nov 25 17:08:23 crc kubenswrapper[4802]: I1125 17:08:23.986865 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-wc8x4"] Nov 25 17:08:24 crc kubenswrapper[4802]: I1125 17:08:24.248768 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 17:08:24 crc kubenswrapper[4802]: I1125 17:08:24.249205 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 17:08:24 crc kubenswrapper[4802]: I1125 17:08:24.955858 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-wc8x4" event={"ID":"2dde71ad-bcda-407e-a490-e8323685e7eb","Type":"ContainerStarted","Data":"6753f6c24486761b31143a299647d5da36bb268414c2eb2817bfaa3ff1139d8e"} Nov 25 17:08:24 crc kubenswrapper[4802]: I1125 17:08:24.955907 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-wc8x4" event={"ID":"2dde71ad-bcda-407e-a490-e8323685e7eb","Type":"ContainerStarted","Data":"d69862b24a168ab04b98df4babf35b9e009bcf75fc2b876681554dedab43f1ea"} Nov 25 17:08:24 crc kubenswrapper[4802]: I1125 17:08:24.976227 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-wc8x4" podStartSLOduration=1.9762043390000001 podStartE2EDuration="1.976204339s" podCreationTimestamp="2025-11-25 17:08:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:08:24.974993357 +0000 UTC m=+1288.119340553" watchObservedRunningTime="2025-11-25 17:08:24.976204339 +0000 UTC m=+1288.120551525" Nov 25 17:08:27 crc kubenswrapper[4802]: I1125 17:08:27.990804 4802 generic.go:334] "Generic (PLEG): container finished" podID="2dde71ad-bcda-407e-a490-e8323685e7eb" containerID="6753f6c24486761b31143a299647d5da36bb268414c2eb2817bfaa3ff1139d8e" exitCode=0 Nov 25 17:08:27 crc kubenswrapper[4802]: I1125 17:08:27.990882 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-wc8x4" event={"ID":"2dde71ad-bcda-407e-a490-e8323685e7eb","Type":"ContainerDied","Data":"6753f6c24486761b31143a299647d5da36bb268414c2eb2817bfaa3ff1139d8e"} Nov 25 17:08:29 crc kubenswrapper[4802]: I1125 17:08:29.369195 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-wc8x4" Nov 25 17:08:29 crc kubenswrapper[4802]: I1125 17:08:29.485409 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2dde71ad-bcda-407e-a490-e8323685e7eb-db-sync-config-data\") pod \"2dde71ad-bcda-407e-a490-e8323685e7eb\" (UID: \"2dde71ad-bcda-407e-a490-e8323685e7eb\") " Nov 25 17:08:29 crc kubenswrapper[4802]: I1125 17:08:29.485713 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ln87\" (UniqueName: \"kubernetes.io/projected/2dde71ad-bcda-407e-a490-e8323685e7eb-kube-api-access-6ln87\") pod \"2dde71ad-bcda-407e-a490-e8323685e7eb\" (UID: \"2dde71ad-bcda-407e-a490-e8323685e7eb\") " Nov 25 17:08:29 crc kubenswrapper[4802]: I1125 17:08:29.486018 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dde71ad-bcda-407e-a490-e8323685e7eb-config-data\") pod \"2dde71ad-bcda-407e-a490-e8323685e7eb\" (UID: \"2dde71ad-bcda-407e-a490-e8323685e7eb\") " Nov 25 17:08:29 crc kubenswrapper[4802]: I1125 17:08:29.492579 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dde71ad-bcda-407e-a490-e8323685e7eb-kube-api-access-6ln87" (OuterVolumeSpecName: "kube-api-access-6ln87") pod "2dde71ad-bcda-407e-a490-e8323685e7eb" (UID: "2dde71ad-bcda-407e-a490-e8323685e7eb"). InnerVolumeSpecName "kube-api-access-6ln87". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:08:29 crc kubenswrapper[4802]: I1125 17:08:29.493401 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dde71ad-bcda-407e-a490-e8323685e7eb-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2dde71ad-bcda-407e-a490-e8323685e7eb" (UID: "2dde71ad-bcda-407e-a490-e8323685e7eb"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:08:29 crc kubenswrapper[4802]: I1125 17:08:29.550612 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dde71ad-bcda-407e-a490-e8323685e7eb-config-data" (OuterVolumeSpecName: "config-data") pod "2dde71ad-bcda-407e-a490-e8323685e7eb" (UID: "2dde71ad-bcda-407e-a490-e8323685e7eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:08:29 crc kubenswrapper[4802]: I1125 17:08:29.588280 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dde71ad-bcda-407e-a490-e8323685e7eb-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:29 crc kubenswrapper[4802]: I1125 17:08:29.588322 4802 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2dde71ad-bcda-407e-a490-e8323685e7eb-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:29 crc kubenswrapper[4802]: I1125 17:08:29.588342 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ln87\" (UniqueName: \"kubernetes.io/projected/2dde71ad-bcda-407e-a490-e8323685e7eb-kube-api-access-6ln87\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:30 crc kubenswrapper[4802]: I1125 17:08:30.008089 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-wc8x4" event={"ID":"2dde71ad-bcda-407e-a490-e8323685e7eb","Type":"ContainerDied","Data":"d69862b24a168ab04b98df4babf35b9e009bcf75fc2b876681554dedab43f1ea"} Nov 25 17:08:30 crc kubenswrapper[4802]: I1125 17:08:30.008137 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-wc8x4" Nov 25 17:08:30 crc kubenswrapper[4802]: I1125 17:08:30.008149 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d69862b24a168ab04b98df4babf35b9e009bcf75fc2b876681554dedab43f1ea" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.241287 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 17:08:31 crc kubenswrapper[4802]: E1125 17:08:31.241746 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dde71ad-bcda-407e-a490-e8323685e7eb" containerName="glance-db-sync" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.241761 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dde71ad-bcda-407e-a490-e8323685e7eb" containerName="glance-db-sync" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.241913 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dde71ad-bcda-407e-a490-e8323685e7eb" containerName="glance-db-sync" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.242778 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.245164 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.245394 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-87pxg" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.245908 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.267338 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.315058 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-sys\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.315105 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.315145 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.315169 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.315193 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-run\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.315218 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6wtz\" (UniqueName: \"kubernetes.io/projected/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-kube-api-access-q6wtz\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.315471 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-dev\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.315549 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.315574 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.315650 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-scripts\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.315698 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-logs\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.315787 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-config-data\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.315807 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.315823 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417016 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-logs\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417086 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-config-data\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417106 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417167 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417202 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-sys\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417213 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417260 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417264 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-sys\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417281 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417301 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417325 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-run\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417349 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6wtz\" (UniqueName: \"kubernetes.io/projected/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-kube-api-access-q6wtz\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417391 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-dev\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417381 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-run\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417350 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417412 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417396 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417425 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-dev\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417534 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417572 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-scripts\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417598 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-logs\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417605 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417637 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.417677 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.418053 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.422387 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-scripts\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.424424 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-config-data\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.435081 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.438859 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.445451 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6wtz\" (UniqueName: \"kubernetes.io/projected/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-kube-api-access-q6wtz\") pod \"glance-default-external-api-0\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.452513 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.454199 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.456420 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.479451 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.527331 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.527413 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.527456 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.527496 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.527529 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.527589 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.527622 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-dev\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.527662 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.527687 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.527724 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-run\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.527774 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-sys\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.527826 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5msm\" (UniqueName: \"kubernetes.io/projected/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-kube-api-access-j5msm\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.527853 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-logs\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.527880 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.574706 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.628935 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5msm\" (UniqueName: \"kubernetes.io/projected/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-kube-api-access-j5msm\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.628986 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-logs\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.629007 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.629028 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.629055 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.629079 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.629103 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.629150 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.629223 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.629240 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-dev\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.629263 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.629278 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.629302 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-run\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.629368 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-sys\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.629450 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-sys\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.629773 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.629924 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.630245 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.630275 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.630248 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-logs\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.630285 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.630368 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-dev\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.630367 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.630324 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-run\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.630433 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.633734 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.637114 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.653959 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.658081 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5msm\" (UniqueName: \"kubernetes.io/projected/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-kube-api-access-j5msm\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.663652 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:31 crc kubenswrapper[4802]: I1125 17:08:31.801923 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:32 crc kubenswrapper[4802]: I1125 17:08:32.009010 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 17:08:32 crc kubenswrapper[4802]: I1125 17:08:32.163364 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 17:08:32 crc kubenswrapper[4802]: I1125 17:08:32.211820 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 17:08:32 crc kubenswrapper[4802]: W1125 17:08:32.218497 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5c481cd_f8af_4de0_8244_be3d73bcbfc6.slice/crio-aeb8b0e109722833e2a1c7884232c9e7639cdf00e7808b037367a1dbd809310d WatchSource:0}: Error finding container aeb8b0e109722833e2a1c7884232c9e7639cdf00e7808b037367a1dbd809310d: Status 404 returned error can't find the container with id aeb8b0e109722833e2a1c7884232c9e7639cdf00e7808b037367a1dbd809310d Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.037612 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b5c481cd-f8af-4de0-8244-be3d73bcbfc6","Type":"ContainerStarted","Data":"454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591"} Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.038227 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b5c481cd-f8af-4de0-8244-be3d73bcbfc6","Type":"ContainerStarted","Data":"5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f"} Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.038251 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b5c481cd-f8af-4de0-8244-be3d73bcbfc6","Type":"ContainerStarted","Data":"80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f"} Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.037747 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="b5c481cd-f8af-4de0-8244-be3d73bcbfc6" containerName="glance-log" containerID="cri-o://80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f" gracePeriod=30 Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.038265 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b5c481cd-f8af-4de0-8244-be3d73bcbfc6","Type":"ContainerStarted","Data":"aeb8b0e109722833e2a1c7884232c9e7639cdf00e7808b037367a1dbd809310d"} Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.037968 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="b5c481cd-f8af-4de0-8244-be3d73bcbfc6" containerName="glance-api" containerID="cri-o://454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591" gracePeriod=30 Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.037889 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="b5c481cd-f8af-4de0-8244-be3d73bcbfc6" containerName="glance-httpd" containerID="cri-o://5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f" gracePeriod=30 Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.045700 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8ccfbe24-ca2d-447d-9675-6836cd5b02b3","Type":"ContainerStarted","Data":"b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051"} Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.045808 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8ccfbe24-ca2d-447d-9675-6836cd5b02b3","Type":"ContainerStarted","Data":"8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e"} Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.045875 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8ccfbe24-ca2d-447d-9675-6836cd5b02b3","Type":"ContainerStarted","Data":"e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b"} Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.045896 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8ccfbe24-ca2d-447d-9675-6836cd5b02b3","Type":"ContainerStarted","Data":"9aed5f3020cd19bf2b496b3f16cb1288a9376c6bd50fcae09eac6ea1631eb731"} Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.083162 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=3.083137628 podStartE2EDuration="3.083137628s" podCreationTimestamp="2025-11-25 17:08:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:08:33.079168542 +0000 UTC m=+1296.223515748" watchObservedRunningTime="2025-11-25 17:08:33.083137628 +0000 UTC m=+1296.227484814" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.115030 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=2.115006956 podStartE2EDuration="2.115006956s" podCreationTimestamp="2025-11-25 17:08:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:08:33.106241121 +0000 UTC m=+1296.250588337" watchObservedRunningTime="2025-11-25 17:08:33.115006956 +0000 UTC m=+1296.259354152" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.496687 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.557845 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-httpd-run\") pod \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.557922 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-etc-iscsi\") pod \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.557964 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.557990 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558037 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-config-data\") pod \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558063 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "b5c481cd-f8af-4de0-8244-be3d73bcbfc6" (UID: "b5c481cd-f8af-4de0-8244-be3d73bcbfc6"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558078 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-lib-modules\") pod \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558161 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "b5c481cd-f8af-4de0-8244-be3d73bcbfc6" (UID: "b5c481cd-f8af-4de0-8244-be3d73bcbfc6"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558194 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-run\") pod \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558228 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-dev\") pod \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558290 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-run" (OuterVolumeSpecName: "run") pod "b5c481cd-f8af-4de0-8244-be3d73bcbfc6" (UID: "b5c481cd-f8af-4de0-8244-be3d73bcbfc6"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558315 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-etc-nvme\") pod \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558331 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-dev" (OuterVolumeSpecName: "dev") pod "b5c481cd-f8af-4de0-8244-be3d73bcbfc6" (UID: "b5c481cd-f8af-4de0-8244-be3d73bcbfc6"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558348 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5msm\" (UniqueName: \"kubernetes.io/projected/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-kube-api-access-j5msm\") pod \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558366 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-sys\") pod \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558383 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-scripts\") pod \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558385 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b5c481cd-f8af-4de0-8244-be3d73bcbfc6" (UID: "b5c481cd-f8af-4de0-8244-be3d73bcbfc6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558398 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "b5c481cd-f8af-4de0-8244-be3d73bcbfc6" (UID: "b5c481cd-f8af-4de0-8244-be3d73bcbfc6"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558403 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-var-locks-brick\") pod \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558483 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-logs\") pod \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\" (UID: \"b5c481cd-f8af-4de0-8244-be3d73bcbfc6\") " Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558422 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "b5c481cd-f8af-4de0-8244-be3d73bcbfc6" (UID: "b5c481cd-f8af-4de0-8244-be3d73bcbfc6"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558445 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-sys" (OuterVolumeSpecName: "sys") pod "b5c481cd-f8af-4de0-8244-be3d73bcbfc6" (UID: "b5c481cd-f8af-4de0-8244-be3d73bcbfc6"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.558797 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-logs" (OuterVolumeSpecName: "logs") pod "b5c481cd-f8af-4de0-8244-be3d73bcbfc6" (UID: "b5c481cd-f8af-4de0-8244-be3d73bcbfc6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.559032 4802 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.559050 4802 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-dev\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.559063 4802 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.559076 4802 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-sys\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.559088 4802 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.559100 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-logs\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.559111 4802 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.559145 4802 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.559162 4802 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.568352 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-scripts" (OuterVolumeSpecName: "scripts") pod "b5c481cd-f8af-4de0-8244-be3d73bcbfc6" (UID: "b5c481cd-f8af-4de0-8244-be3d73bcbfc6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.568413 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-kube-api-access-j5msm" (OuterVolumeSpecName: "kube-api-access-j5msm") pod "b5c481cd-f8af-4de0-8244-be3d73bcbfc6" (UID: "b5c481cd-f8af-4de0-8244-be3d73bcbfc6"). InnerVolumeSpecName "kube-api-access-j5msm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.568440 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "b5c481cd-f8af-4de0-8244-be3d73bcbfc6" (UID: "b5c481cd-f8af-4de0-8244-be3d73bcbfc6"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.568434 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance-cache") pod "b5c481cd-f8af-4de0-8244-be3d73bcbfc6" (UID: "b5c481cd-f8af-4de0-8244-be3d73bcbfc6"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.653257 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-config-data" (OuterVolumeSpecName: "config-data") pod "b5c481cd-f8af-4de0-8244-be3d73bcbfc6" (UID: "b5c481cd-f8af-4de0-8244-be3d73bcbfc6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.660864 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5msm\" (UniqueName: \"kubernetes.io/projected/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-kube-api-access-j5msm\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.660894 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.660929 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.660942 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.660952 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5c481cd-f8af-4de0-8244-be3d73bcbfc6-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.678680 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.685954 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.762350 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:33 crc kubenswrapper[4802]: I1125 17:08:33.762402 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.055645 4802 generic.go:334] "Generic (PLEG): container finished" podID="b5c481cd-f8af-4de0-8244-be3d73bcbfc6" containerID="454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591" exitCode=143 Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.055688 4802 generic.go:334] "Generic (PLEG): container finished" podID="b5c481cd-f8af-4de0-8244-be3d73bcbfc6" containerID="5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f" exitCode=143 Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.055705 4802 generic.go:334] "Generic (PLEG): container finished" podID="b5c481cd-f8af-4de0-8244-be3d73bcbfc6" containerID="80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f" exitCode=143 Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.055738 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.055783 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b5c481cd-f8af-4de0-8244-be3d73bcbfc6","Type":"ContainerDied","Data":"454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591"} Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.055839 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b5c481cd-f8af-4de0-8244-be3d73bcbfc6","Type":"ContainerDied","Data":"5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f"} Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.055856 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b5c481cd-f8af-4de0-8244-be3d73bcbfc6","Type":"ContainerDied","Data":"80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f"} Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.055866 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b5c481cd-f8af-4de0-8244-be3d73bcbfc6","Type":"ContainerDied","Data":"aeb8b0e109722833e2a1c7884232c9e7639cdf00e7808b037367a1dbd809310d"} Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.055882 4802 scope.go:117] "RemoveContainer" containerID="454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.089735 4802 scope.go:117] "RemoveContainer" containerID="5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.099820 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.126778 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.141716 4802 scope.go:117] "RemoveContainer" containerID="80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.143315 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 17:08:34 crc kubenswrapper[4802]: E1125 17:08:34.143763 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5c481cd-f8af-4de0-8244-be3d73bcbfc6" containerName="glance-log" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.143786 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5c481cd-f8af-4de0-8244-be3d73bcbfc6" containerName="glance-log" Nov 25 17:08:34 crc kubenswrapper[4802]: E1125 17:08:34.143828 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5c481cd-f8af-4de0-8244-be3d73bcbfc6" containerName="glance-api" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.143839 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5c481cd-f8af-4de0-8244-be3d73bcbfc6" containerName="glance-api" Nov 25 17:08:34 crc kubenswrapper[4802]: E1125 17:08:34.143873 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5c481cd-f8af-4de0-8244-be3d73bcbfc6" containerName="glance-httpd" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.143884 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5c481cd-f8af-4de0-8244-be3d73bcbfc6" containerName="glance-httpd" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.144418 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5c481cd-f8af-4de0-8244-be3d73bcbfc6" containerName="glance-api" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.144448 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5c481cd-f8af-4de0-8244-be3d73bcbfc6" containerName="glance-log" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.144477 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5c481cd-f8af-4de0-8244-be3d73bcbfc6" containerName="glance-httpd" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.145894 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.148498 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.149831 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.177043 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f9ba153-fd78-4203-b36f-39e14d80e747-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.177165 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.177209 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.177901 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f9ba153-fd78-4203-b36f-39e14d80e747-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.177992 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-dev\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.178025 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.178143 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f9ba153-fd78-4203-b36f-39e14d80e747-logs\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.178189 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.178218 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f9ba153-fd78-4203-b36f-39e14d80e747-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.178260 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-sys\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.178326 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.178373 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjvzp\" (UniqueName: \"kubernetes.io/projected/1f9ba153-fd78-4203-b36f-39e14d80e747-kube-api-access-mjvzp\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.179221 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.179275 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-run\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.181462 4802 scope.go:117] "RemoveContainer" containerID="454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591" Nov 25 17:08:34 crc kubenswrapper[4802]: E1125 17:08:34.182237 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591\": container with ID starting with 454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591 not found: ID does not exist" containerID="454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.182271 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591"} err="failed to get container status \"454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591\": rpc error: code = NotFound desc = could not find container \"454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591\": container with ID starting with 454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591 not found: ID does not exist" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.182299 4802 scope.go:117] "RemoveContainer" containerID="5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f" Nov 25 17:08:34 crc kubenswrapper[4802]: E1125 17:08:34.182801 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f\": container with ID starting with 5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f not found: ID does not exist" containerID="5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.182821 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f"} err="failed to get container status \"5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f\": rpc error: code = NotFound desc = could not find container \"5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f\": container with ID starting with 5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f not found: ID does not exist" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.182835 4802 scope.go:117] "RemoveContainer" containerID="80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f" Nov 25 17:08:34 crc kubenswrapper[4802]: E1125 17:08:34.183108 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f\": container with ID starting with 80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f not found: ID does not exist" containerID="80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.183157 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f"} err="failed to get container status \"80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f\": rpc error: code = NotFound desc = could not find container \"80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f\": container with ID starting with 80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f not found: ID does not exist" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.183186 4802 scope.go:117] "RemoveContainer" containerID="454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.183520 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591"} err="failed to get container status \"454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591\": rpc error: code = NotFound desc = could not find container \"454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591\": container with ID starting with 454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591 not found: ID does not exist" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.183543 4802 scope.go:117] "RemoveContainer" containerID="5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.183887 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f"} err="failed to get container status \"5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f\": rpc error: code = NotFound desc = could not find container \"5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f\": container with ID starting with 5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f not found: ID does not exist" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.183904 4802 scope.go:117] "RemoveContainer" containerID="80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.184193 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f"} err="failed to get container status \"80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f\": rpc error: code = NotFound desc = could not find container \"80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f\": container with ID starting with 80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f not found: ID does not exist" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.184210 4802 scope.go:117] "RemoveContainer" containerID="454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.184521 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591"} err="failed to get container status \"454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591\": rpc error: code = NotFound desc = could not find container \"454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591\": container with ID starting with 454c3b763adae31933a597fd0026ddf923a0b2784cee6d64f2658542c89a8591 not found: ID does not exist" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.184597 4802 scope.go:117] "RemoveContainer" containerID="5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.185344 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f"} err="failed to get container status \"5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f\": rpc error: code = NotFound desc = could not find container \"5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f\": container with ID starting with 5ba4213bc9527ddef40de5974e74db554990aa942559b2f70ff0f99aca0ead7f not found: ID does not exist" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.185368 4802 scope.go:117] "RemoveContainer" containerID="80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.185960 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f"} err="failed to get container status \"80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f\": rpc error: code = NotFound desc = could not find container \"80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f\": container with ID starting with 80a24ac2248302bd98a6135428ccfe17008dce99af162c3eda0895dce6ee869f not found: ID does not exist" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.280615 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f9ba153-fd78-4203-b36f-39e14d80e747-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.280712 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-sys\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.280781 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.281034 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjvzp\" (UniqueName: \"kubernetes.io/projected/1f9ba153-fd78-4203-b36f-39e14d80e747-kube-api-access-mjvzp\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.281105 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.281228 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-run\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.281275 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f9ba153-fd78-4203-b36f-39e14d80e747-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.281337 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.281401 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.281465 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f9ba153-fd78-4203-b36f-39e14d80e747-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.281553 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-dev\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.281603 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.281681 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f9ba153-fd78-4203-b36f-39e14d80e747-logs\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.281734 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.282212 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.282917 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.282994 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-dev\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.283063 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.283058 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.283110 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-sys\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.283867 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-run\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.284020 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f9ba153-fd78-4203-b36f-39e14d80e747-logs\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.284217 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f9ba153-fd78-4203-b36f-39e14d80e747-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.284308 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.283377 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.290274 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f9ba153-fd78-4203-b36f-39e14d80e747-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.301989 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f9ba153-fd78-4203-b36f-39e14d80e747-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.317111 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.317797 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjvzp\" (UniqueName: \"kubernetes.io/projected/1f9ba153-fd78-4203-b36f-39e14d80e747-kube-api-access-mjvzp\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.581632 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:34 crc kubenswrapper[4802]: I1125 17:08:34.773837 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:35 crc kubenswrapper[4802]: I1125 17:08:35.261921 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 17:08:35 crc kubenswrapper[4802]: W1125 17:08:35.269719 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f9ba153_fd78_4203_b36f_39e14d80e747.slice/crio-6c25a06e246e174a57dd57e540afd10843aa870584b2aa2a5d4dbec3b20dbb55 WatchSource:0}: Error finding container 6c25a06e246e174a57dd57e540afd10843aa870584b2aa2a5d4dbec3b20dbb55: Status 404 returned error can't find the container with id 6c25a06e246e174a57dd57e540afd10843aa870584b2aa2a5d4dbec3b20dbb55 Nov 25 17:08:35 crc kubenswrapper[4802]: I1125 17:08:35.514203 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5c481cd-f8af-4de0-8244-be3d73bcbfc6" path="/var/lib/kubelet/pods/b5c481cd-f8af-4de0-8244-be3d73bcbfc6/volumes" Nov 25 17:08:36 crc kubenswrapper[4802]: I1125 17:08:36.084070 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1f9ba153-fd78-4203-b36f-39e14d80e747","Type":"ContainerStarted","Data":"27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6"} Nov 25 17:08:36 crc kubenswrapper[4802]: I1125 17:08:36.084464 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1f9ba153-fd78-4203-b36f-39e14d80e747","Type":"ContainerStarted","Data":"caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca"} Nov 25 17:08:36 crc kubenswrapper[4802]: I1125 17:08:36.084501 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1f9ba153-fd78-4203-b36f-39e14d80e747","Type":"ContainerStarted","Data":"fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef"} Nov 25 17:08:36 crc kubenswrapper[4802]: I1125 17:08:36.084513 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1f9ba153-fd78-4203-b36f-39e14d80e747","Type":"ContainerStarted","Data":"6c25a06e246e174a57dd57e540afd10843aa870584b2aa2a5d4dbec3b20dbb55"} Nov 25 17:08:36 crc kubenswrapper[4802]: I1125 17:08:36.124952 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.124909975 podStartE2EDuration="2.124909975s" podCreationTimestamp="2025-11-25 17:08:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:08:36.117647739 +0000 UTC m=+1299.261994945" watchObservedRunningTime="2025-11-25 17:08:36.124909975 +0000 UTC m=+1299.269257161" Nov 25 17:08:41 crc kubenswrapper[4802]: I1125 17:08:41.575825 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:41 crc kubenswrapper[4802]: I1125 17:08:41.576517 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:41 crc kubenswrapper[4802]: I1125 17:08:41.576545 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:41 crc kubenswrapper[4802]: I1125 17:08:41.618359 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:41 crc kubenswrapper[4802]: I1125 17:08:41.618928 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:41 crc kubenswrapper[4802]: I1125 17:08:41.631576 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:42 crc kubenswrapper[4802]: I1125 17:08:42.140013 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:42 crc kubenswrapper[4802]: I1125 17:08:42.140366 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:42 crc kubenswrapper[4802]: I1125 17:08:42.140469 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:42 crc kubenswrapper[4802]: I1125 17:08:42.152203 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:42 crc kubenswrapper[4802]: I1125 17:08:42.152861 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:42 crc kubenswrapper[4802]: I1125 17:08:42.154671 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:08:44 crc kubenswrapper[4802]: I1125 17:08:44.774109 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:44 crc kubenswrapper[4802]: I1125 17:08:44.774445 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:44 crc kubenswrapper[4802]: I1125 17:08:44.774460 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:44 crc kubenswrapper[4802]: I1125 17:08:44.799494 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:44 crc kubenswrapper[4802]: I1125 17:08:44.806169 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:44 crc kubenswrapper[4802]: I1125 17:08:44.824330 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:45 crc kubenswrapper[4802]: I1125 17:08:45.161651 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:45 crc kubenswrapper[4802]: I1125 17:08:45.162028 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:45 crc kubenswrapper[4802]: I1125 17:08:45.162284 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:45 crc kubenswrapper[4802]: I1125 17:08:45.174907 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:45 crc kubenswrapper[4802]: I1125 17:08:45.182196 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:45 crc kubenswrapper[4802]: I1125 17:08:45.183193 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.484723 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.486780 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.492509 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.499402 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.501590 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.522661 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.598067 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.598145 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.598173 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.598219 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-run\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.598241 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/723777a0-7ac9-42ee-9d5e-95175c0272f9-logs\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.598284 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.598310 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.598331 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4f07881-4b13-440d-b080-25f1a5df75ac-scripts\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.598744 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/723777a0-7ac9-42ee-9d5e-95175c0272f9-config-data\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.598933 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-run\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.599055 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4f07881-4b13-440d-b080-25f1a5df75ac-config-data\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.599094 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggxt2\" (UniqueName: \"kubernetes.io/projected/723777a0-7ac9-42ee-9d5e-95175c0272f9-kube-api-access-ggxt2\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.599188 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/723777a0-7ac9-42ee-9d5e-95175c0272f9-scripts\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.599241 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-dev\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.599305 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4f07881-4b13-440d-b080-25f1a5df75ac-logs\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.599364 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fzlp\" (UniqueName: \"kubernetes.io/projected/a4f07881-4b13-440d-b080-25f1a5df75ac-kube-api-access-5fzlp\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.599425 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-sys\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.599492 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.599526 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-sys\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.599611 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.599662 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.599705 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-dev\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.599735 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.599789 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.599812 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.599832 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.599858 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/723777a0-7ac9-42ee-9d5e-95175c0272f9-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.599881 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a4f07881-4b13-440d-b080-25f1a5df75ac-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701631 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4f07881-4b13-440d-b080-25f1a5df75ac-config-data\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701686 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggxt2\" (UniqueName: \"kubernetes.io/projected/723777a0-7ac9-42ee-9d5e-95175c0272f9-kube-api-access-ggxt2\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701710 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/723777a0-7ac9-42ee-9d5e-95175c0272f9-scripts\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701727 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-dev\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701749 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4f07881-4b13-440d-b080-25f1a5df75ac-logs\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701770 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fzlp\" (UniqueName: \"kubernetes.io/projected/a4f07881-4b13-440d-b080-25f1a5df75ac-kube-api-access-5fzlp\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701788 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-sys\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701808 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701822 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-sys\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701837 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701854 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701869 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-dev\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701887 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701904 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701917 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701933 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a4f07881-4b13-440d-b080-25f1a5df75ac-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701948 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-sys\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701983 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.701946 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702011 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702032 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-sys\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702036 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/723777a0-7ac9-42ee-9d5e-95175c0272f9-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702093 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702203 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702252 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702342 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702375 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702530 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-dev\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702534 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702542 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702542 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702589 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702535 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-dev\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702589 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702625 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702815 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-run\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702873 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/723777a0-7ac9-42ee-9d5e-95175c0272f9-logs\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702911 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-run\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.702886 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a4f07881-4b13-440d-b080-25f1a5df75ac-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.703284 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.703360 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/723777a0-7ac9-42ee-9d5e-95175c0272f9-logs\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.703371 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.703427 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4f07881-4b13-440d-b080-25f1a5df75ac-scripts\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.703469 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4f07881-4b13-440d-b080-25f1a5df75ac-logs\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.703484 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/723777a0-7ac9-42ee-9d5e-95175c0272f9-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.703585 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.703592 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.703791 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/723777a0-7ac9-42ee-9d5e-95175c0272f9-config-data\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.703850 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-run\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.704020 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-run\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.710922 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4f07881-4b13-440d-b080-25f1a5df75ac-config-data\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.711171 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/723777a0-7ac9-42ee-9d5e-95175c0272f9-scripts\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.715977 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4f07881-4b13-440d-b080-25f1a5df75ac-scripts\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.717061 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/723777a0-7ac9-42ee-9d5e-95175c0272f9-config-data\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.720568 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggxt2\" (UniqueName: \"kubernetes.io/projected/723777a0-7ac9-42ee-9d5e-95175c0272f9-kube-api-access-ggxt2\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.728829 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fzlp\" (UniqueName: \"kubernetes.io/projected/a4f07881-4b13-440d-b080-25f1a5df75ac-kube-api-access-5fzlp\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.741343 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.741919 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-1\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.757883 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.758554 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-2\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.823455 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:47 crc kubenswrapper[4802]: I1125 17:08:47.839945 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:48 crc kubenswrapper[4802]: I1125 17:08:48.056696 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 17:08:48 crc kubenswrapper[4802]: I1125 17:08:48.093073 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 17:08:48 crc kubenswrapper[4802]: W1125 17:08:48.101980 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4f07881_4b13_440d_b080_25f1a5df75ac.slice/crio-391c01c5b7b031135ea27c5b222f13a7bb0e6a803dcc805821cbde6f9f8de399 WatchSource:0}: Error finding container 391c01c5b7b031135ea27c5b222f13a7bb0e6a803dcc805821cbde6f9f8de399: Status 404 returned error can't find the container with id 391c01c5b7b031135ea27c5b222f13a7bb0e6a803dcc805821cbde6f9f8de399 Nov 25 17:08:48 crc kubenswrapper[4802]: I1125 17:08:48.181410 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"a4f07881-4b13-440d-b080-25f1a5df75ac","Type":"ContainerStarted","Data":"391c01c5b7b031135ea27c5b222f13a7bb0e6a803dcc805821cbde6f9f8de399"} Nov 25 17:08:48 crc kubenswrapper[4802]: I1125 17:08:48.182354 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"723777a0-7ac9-42ee-9d5e-95175c0272f9","Type":"ContainerStarted","Data":"b9a543ebc7d3a78a725ff30ba52f56b6569adbec8da2f593758a4674ce73330f"} Nov 25 17:08:49 crc kubenswrapper[4802]: I1125 17:08:49.197605 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"723777a0-7ac9-42ee-9d5e-95175c0272f9","Type":"ContainerStarted","Data":"8604befb4306036f25199f027d8899febaa5dabc3aba554146758c29205052ed"} Nov 25 17:08:49 crc kubenswrapper[4802]: I1125 17:08:49.198238 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"723777a0-7ac9-42ee-9d5e-95175c0272f9","Type":"ContainerStarted","Data":"0898fe2b1471d347c1fc6462e60173cb851ab29582cf64c0fdd4629e353bef48"} Nov 25 17:08:49 crc kubenswrapper[4802]: I1125 17:08:49.198257 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"723777a0-7ac9-42ee-9d5e-95175c0272f9","Type":"ContainerStarted","Data":"5b68c1225cb655d3c617515bc47a9a58a5fc3b1c5e871c98dc38166d1ef8a4e9"} Nov 25 17:08:49 crc kubenswrapper[4802]: I1125 17:08:49.205284 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"a4f07881-4b13-440d-b080-25f1a5df75ac","Type":"ContainerStarted","Data":"69b82248fd43aafc0103f6112021d7f1e2e8a14f6332a08cd5e445eee67e29e6"} Nov 25 17:08:49 crc kubenswrapper[4802]: I1125 17:08:49.205321 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"a4f07881-4b13-440d-b080-25f1a5df75ac","Type":"ContainerStarted","Data":"7469712d663f353b3939791a9bb5166b011d571df689064f7849ce4c55a0108a"} Nov 25 17:08:49 crc kubenswrapper[4802]: I1125 17:08:49.205335 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"a4f07881-4b13-440d-b080-25f1a5df75ac","Type":"ContainerStarted","Data":"36f189eb936f1203af5299dde514f6790364ae557f8cb3c191c3c8f93b1b3fa3"} Nov 25 17:08:49 crc kubenswrapper[4802]: I1125 17:08:49.247940 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=3.247918586 podStartE2EDuration="3.247918586s" podCreationTimestamp="2025-11-25 17:08:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:08:49.236445106 +0000 UTC m=+1312.380792292" watchObservedRunningTime="2025-11-25 17:08:49.247918586 +0000 UTC m=+1312.392265772" Nov 25 17:08:49 crc kubenswrapper[4802]: I1125 17:08:49.280270 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-2" podStartSLOduration=3.2802342859999998 podStartE2EDuration="3.280234286s" podCreationTimestamp="2025-11-25 17:08:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:08:49.272663521 +0000 UTC m=+1312.417010737" watchObservedRunningTime="2025-11-25 17:08:49.280234286 +0000 UTC m=+1312.424581512" Nov 25 17:08:54 crc kubenswrapper[4802]: I1125 17:08:54.248728 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 17:08:54 crc kubenswrapper[4802]: I1125 17:08:54.249249 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 17:08:57 crc kubenswrapper[4802]: I1125 17:08:57.825792 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:57 crc kubenswrapper[4802]: I1125 17:08:57.826199 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:57 crc kubenswrapper[4802]: I1125 17:08:57.826265 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:57 crc kubenswrapper[4802]: I1125 17:08:57.847022 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:57 crc kubenswrapper[4802]: I1125 17:08:57.847116 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:57 crc kubenswrapper[4802]: I1125 17:08:57.847196 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:57 crc kubenswrapper[4802]: I1125 17:08:57.864262 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:57 crc kubenswrapper[4802]: I1125 17:08:57.877342 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:57 crc kubenswrapper[4802]: I1125 17:08:57.887188 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:57 crc kubenswrapper[4802]: I1125 17:08:57.891243 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:57 crc kubenswrapper[4802]: I1125 17:08:57.892324 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:57 crc kubenswrapper[4802]: I1125 17:08:57.906287 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:58 crc kubenswrapper[4802]: I1125 17:08:58.284582 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:58 crc kubenswrapper[4802]: I1125 17:08:58.284616 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:58 crc kubenswrapper[4802]: I1125 17:08:58.284626 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:58 crc kubenswrapper[4802]: I1125 17:08:58.284635 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:58 crc kubenswrapper[4802]: I1125 17:08:58.284644 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:58 crc kubenswrapper[4802]: I1125 17:08:58.284653 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:58 crc kubenswrapper[4802]: I1125 17:08:58.307363 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:58 crc kubenswrapper[4802]: I1125 17:08:58.309672 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:58 crc kubenswrapper[4802]: I1125 17:08:58.310458 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:08:58 crc kubenswrapper[4802]: I1125 17:08:58.391661 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:58 crc kubenswrapper[4802]: I1125 17:08:58.416606 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:08:58 crc kubenswrapper[4802]: I1125 17:08:58.453259 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:09:24 crc kubenswrapper[4802]: I1125 17:09:24.260095 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 17:09:24 crc kubenswrapper[4802]: I1125 17:09:24.260994 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 17:09:24 crc kubenswrapper[4802]: I1125 17:09:24.261074 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 17:09:24 crc kubenswrapper[4802]: I1125 17:09:24.262294 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ee1883bc854fb1518fe7b71c8f9bba592e9827245176f86eb7e6da8b81dacf06"} pod="openshift-machine-config-operator/machine-config-daemon-h29wc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 17:09:24 crc kubenswrapper[4802]: I1125 17:09:24.262419 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" containerID="cri-o://ee1883bc854fb1518fe7b71c8f9bba592e9827245176f86eb7e6da8b81dacf06" gracePeriod=600 Nov 25 17:09:24 crc kubenswrapper[4802]: I1125 17:09:24.528633 4802 generic.go:334] "Generic (PLEG): container finished" podID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerID="ee1883bc854fb1518fe7b71c8f9bba592e9827245176f86eb7e6da8b81dacf06" exitCode=0 Nov 25 17:09:24 crc kubenswrapper[4802]: I1125 17:09:24.528730 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerDied","Data":"ee1883bc854fb1518fe7b71c8f9bba592e9827245176f86eb7e6da8b81dacf06"} Nov 25 17:09:24 crc kubenswrapper[4802]: I1125 17:09:24.529096 4802 scope.go:117] "RemoveContainer" containerID="18e081a4af641b516662b68a934da8c320eec5e2196f1740999f260df1353261" Nov 25 17:09:25 crc kubenswrapper[4802]: I1125 17:09:25.542858 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerStarted","Data":"28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29"} Nov 25 17:10:44 crc kubenswrapper[4802]: I1125 17:10:44.259505 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7p7fp"] Nov 25 17:10:44 crc kubenswrapper[4802]: I1125 17:10:44.261722 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7p7fp" Nov 25 17:10:44 crc kubenswrapper[4802]: I1125 17:10:44.270290 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7p7fp"] Nov 25 17:10:44 crc kubenswrapper[4802]: I1125 17:10:44.359611 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5018a6f-306c-4866-b72f-1810b53c92ba-utilities\") pod \"certified-operators-7p7fp\" (UID: \"a5018a6f-306c-4866-b72f-1810b53c92ba\") " pod="openshift-marketplace/certified-operators-7p7fp" Nov 25 17:10:44 crc kubenswrapper[4802]: I1125 17:10:44.359681 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5tr9\" (UniqueName: \"kubernetes.io/projected/a5018a6f-306c-4866-b72f-1810b53c92ba-kube-api-access-q5tr9\") pod \"certified-operators-7p7fp\" (UID: \"a5018a6f-306c-4866-b72f-1810b53c92ba\") " pod="openshift-marketplace/certified-operators-7p7fp" Nov 25 17:10:44 crc kubenswrapper[4802]: I1125 17:10:44.359746 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5018a6f-306c-4866-b72f-1810b53c92ba-catalog-content\") pod \"certified-operators-7p7fp\" (UID: \"a5018a6f-306c-4866-b72f-1810b53c92ba\") " pod="openshift-marketplace/certified-operators-7p7fp" Nov 25 17:10:44 crc kubenswrapper[4802]: I1125 17:10:44.461581 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5018a6f-306c-4866-b72f-1810b53c92ba-utilities\") pod \"certified-operators-7p7fp\" (UID: \"a5018a6f-306c-4866-b72f-1810b53c92ba\") " pod="openshift-marketplace/certified-operators-7p7fp" Nov 25 17:10:44 crc kubenswrapper[4802]: I1125 17:10:44.461671 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5tr9\" (UniqueName: \"kubernetes.io/projected/a5018a6f-306c-4866-b72f-1810b53c92ba-kube-api-access-q5tr9\") pod \"certified-operators-7p7fp\" (UID: \"a5018a6f-306c-4866-b72f-1810b53c92ba\") " pod="openshift-marketplace/certified-operators-7p7fp" Nov 25 17:10:44 crc kubenswrapper[4802]: I1125 17:10:44.461753 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5018a6f-306c-4866-b72f-1810b53c92ba-catalog-content\") pod \"certified-operators-7p7fp\" (UID: \"a5018a6f-306c-4866-b72f-1810b53c92ba\") " pod="openshift-marketplace/certified-operators-7p7fp" Nov 25 17:10:44 crc kubenswrapper[4802]: I1125 17:10:44.462351 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5018a6f-306c-4866-b72f-1810b53c92ba-utilities\") pod \"certified-operators-7p7fp\" (UID: \"a5018a6f-306c-4866-b72f-1810b53c92ba\") " pod="openshift-marketplace/certified-operators-7p7fp" Nov 25 17:10:44 crc kubenswrapper[4802]: I1125 17:10:44.462398 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5018a6f-306c-4866-b72f-1810b53c92ba-catalog-content\") pod \"certified-operators-7p7fp\" (UID: \"a5018a6f-306c-4866-b72f-1810b53c92ba\") " pod="openshift-marketplace/certified-operators-7p7fp" Nov 25 17:10:44 crc kubenswrapper[4802]: I1125 17:10:44.482669 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5tr9\" (UniqueName: \"kubernetes.io/projected/a5018a6f-306c-4866-b72f-1810b53c92ba-kube-api-access-q5tr9\") pod \"certified-operators-7p7fp\" (UID: \"a5018a6f-306c-4866-b72f-1810b53c92ba\") " pod="openshift-marketplace/certified-operators-7p7fp" Nov 25 17:10:44 crc kubenswrapper[4802]: I1125 17:10:44.587977 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7p7fp" Nov 25 17:10:45 crc kubenswrapper[4802]: I1125 17:10:45.036612 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7p7fp"] Nov 25 17:10:45 crc kubenswrapper[4802]: I1125 17:10:45.225202 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p7fp" event={"ID":"a5018a6f-306c-4866-b72f-1810b53c92ba","Type":"ContainerStarted","Data":"4b79f5e4296394aaafc2c8f4680acce04662fa63956036fe7e0c6834bfdeb54d"} Nov 25 17:10:45 crc kubenswrapper[4802]: I1125 17:10:45.225421 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p7fp" event={"ID":"a5018a6f-306c-4866-b72f-1810b53c92ba","Type":"ContainerStarted","Data":"eab5c3179d3fe5241ac1cd3c9c6393a00a27b8fa9b520db9c5da2689e602ad89"} Nov 25 17:10:46 crc kubenswrapper[4802]: I1125 17:10:46.235798 4802 generic.go:334] "Generic (PLEG): container finished" podID="a5018a6f-306c-4866-b72f-1810b53c92ba" containerID="4b79f5e4296394aaafc2c8f4680acce04662fa63956036fe7e0c6834bfdeb54d" exitCode=0 Nov 25 17:10:46 crc kubenswrapper[4802]: I1125 17:10:46.235878 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p7fp" event={"ID":"a5018a6f-306c-4866-b72f-1810b53c92ba","Type":"ContainerDied","Data":"4b79f5e4296394aaafc2c8f4680acce04662fa63956036fe7e0c6834bfdeb54d"} Nov 25 17:10:46 crc kubenswrapper[4802]: I1125 17:10:46.240958 4802 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 17:10:47 crc kubenswrapper[4802]: I1125 17:10:47.248193 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p7fp" event={"ID":"a5018a6f-306c-4866-b72f-1810b53c92ba","Type":"ContainerStarted","Data":"2a8bdd0e31e006cf2fc8217352e56a19ec1c338a7e61e7aa598c1a395f5d0822"} Nov 25 17:10:48 crc kubenswrapper[4802]: I1125 17:10:48.259322 4802 generic.go:334] "Generic (PLEG): container finished" podID="a5018a6f-306c-4866-b72f-1810b53c92ba" containerID="2a8bdd0e31e006cf2fc8217352e56a19ec1c338a7e61e7aa598c1a395f5d0822" exitCode=0 Nov 25 17:10:48 crc kubenswrapper[4802]: I1125 17:10:48.259377 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p7fp" event={"ID":"a5018a6f-306c-4866-b72f-1810b53c92ba","Type":"ContainerDied","Data":"2a8bdd0e31e006cf2fc8217352e56a19ec1c338a7e61e7aa598c1a395f5d0822"} Nov 25 17:10:49 crc kubenswrapper[4802]: I1125 17:10:49.268481 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p7fp" event={"ID":"a5018a6f-306c-4866-b72f-1810b53c92ba","Type":"ContainerStarted","Data":"30254af7c113bab1a2f7709ddb5cae6f1c236242e67aa88d4ea6c0532141dde0"} Nov 25 17:10:49 crc kubenswrapper[4802]: I1125 17:10:49.318165 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7p7fp" podStartSLOduration=2.920286827 podStartE2EDuration="5.318114666s" podCreationTimestamp="2025-11-25 17:10:44 +0000 UTC" firstStartedPulling="2025-11-25 17:10:46.239318883 +0000 UTC m=+1429.383666079" lastFinishedPulling="2025-11-25 17:10:48.637146732 +0000 UTC m=+1431.781493918" observedRunningTime="2025-11-25 17:10:49.316702388 +0000 UTC m=+1432.461049574" watchObservedRunningTime="2025-11-25 17:10:49.318114666 +0000 UTC m=+1432.462461862" Nov 25 17:10:54 crc kubenswrapper[4802]: I1125 17:10:54.588173 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7p7fp" Nov 25 17:10:54 crc kubenswrapper[4802]: I1125 17:10:54.588620 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7p7fp" Nov 25 17:10:54 crc kubenswrapper[4802]: I1125 17:10:54.672921 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7p7fp" Nov 25 17:10:55 crc kubenswrapper[4802]: I1125 17:10:55.361646 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7p7fp" Nov 25 17:10:55 crc kubenswrapper[4802]: I1125 17:10:55.401610 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7p7fp"] Nov 25 17:10:57 crc kubenswrapper[4802]: I1125 17:10:57.338816 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7p7fp" podUID="a5018a6f-306c-4866-b72f-1810b53c92ba" containerName="registry-server" containerID="cri-o://30254af7c113bab1a2f7709ddb5cae6f1c236242e67aa88d4ea6c0532141dde0" gracePeriod=2 Nov 25 17:10:57 crc kubenswrapper[4802]: I1125 17:10:57.738569 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7p7fp" Nov 25 17:10:57 crc kubenswrapper[4802]: I1125 17:10:57.790586 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5018a6f-306c-4866-b72f-1810b53c92ba-utilities\") pod \"a5018a6f-306c-4866-b72f-1810b53c92ba\" (UID: \"a5018a6f-306c-4866-b72f-1810b53c92ba\") " Nov 25 17:10:57 crc kubenswrapper[4802]: I1125 17:10:57.790715 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5tr9\" (UniqueName: \"kubernetes.io/projected/a5018a6f-306c-4866-b72f-1810b53c92ba-kube-api-access-q5tr9\") pod \"a5018a6f-306c-4866-b72f-1810b53c92ba\" (UID: \"a5018a6f-306c-4866-b72f-1810b53c92ba\") " Nov 25 17:10:57 crc kubenswrapper[4802]: I1125 17:10:57.790771 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5018a6f-306c-4866-b72f-1810b53c92ba-catalog-content\") pod \"a5018a6f-306c-4866-b72f-1810b53c92ba\" (UID: \"a5018a6f-306c-4866-b72f-1810b53c92ba\") " Nov 25 17:10:57 crc kubenswrapper[4802]: I1125 17:10:57.791865 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5018a6f-306c-4866-b72f-1810b53c92ba-utilities" (OuterVolumeSpecName: "utilities") pod "a5018a6f-306c-4866-b72f-1810b53c92ba" (UID: "a5018a6f-306c-4866-b72f-1810b53c92ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:10:57 crc kubenswrapper[4802]: I1125 17:10:57.795960 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5018a6f-306c-4866-b72f-1810b53c92ba-kube-api-access-q5tr9" (OuterVolumeSpecName: "kube-api-access-q5tr9") pod "a5018a6f-306c-4866-b72f-1810b53c92ba" (UID: "a5018a6f-306c-4866-b72f-1810b53c92ba"). InnerVolumeSpecName "kube-api-access-q5tr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:10:57 crc kubenswrapper[4802]: I1125 17:10:57.831279 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5018a6f-306c-4866-b72f-1810b53c92ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5018a6f-306c-4866-b72f-1810b53c92ba" (UID: "a5018a6f-306c-4866-b72f-1810b53c92ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:10:57 crc kubenswrapper[4802]: I1125 17:10:57.893264 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5018a6f-306c-4866-b72f-1810b53c92ba-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 17:10:57 crc kubenswrapper[4802]: I1125 17:10:57.893301 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5tr9\" (UniqueName: \"kubernetes.io/projected/a5018a6f-306c-4866-b72f-1810b53c92ba-kube-api-access-q5tr9\") on node \"crc\" DevicePath \"\"" Nov 25 17:10:57 crc kubenswrapper[4802]: I1125 17:10:57.893314 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5018a6f-306c-4866-b72f-1810b53c92ba-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 17:10:58 crc kubenswrapper[4802]: I1125 17:10:58.350954 4802 generic.go:334] "Generic (PLEG): container finished" podID="a5018a6f-306c-4866-b72f-1810b53c92ba" containerID="30254af7c113bab1a2f7709ddb5cae6f1c236242e67aa88d4ea6c0532141dde0" exitCode=0 Nov 25 17:10:58 crc kubenswrapper[4802]: I1125 17:10:58.351038 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p7fp" event={"ID":"a5018a6f-306c-4866-b72f-1810b53c92ba","Type":"ContainerDied","Data":"30254af7c113bab1a2f7709ddb5cae6f1c236242e67aa88d4ea6c0532141dde0"} Nov 25 17:10:58 crc kubenswrapper[4802]: I1125 17:10:58.351062 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7p7fp" Nov 25 17:10:58 crc kubenswrapper[4802]: I1125 17:10:58.351103 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p7fp" event={"ID":"a5018a6f-306c-4866-b72f-1810b53c92ba","Type":"ContainerDied","Data":"eab5c3179d3fe5241ac1cd3c9c6393a00a27b8fa9b520db9c5da2689e602ad89"} Nov 25 17:10:58 crc kubenswrapper[4802]: I1125 17:10:58.351167 4802 scope.go:117] "RemoveContainer" containerID="30254af7c113bab1a2f7709ddb5cae6f1c236242e67aa88d4ea6c0532141dde0" Nov 25 17:10:58 crc kubenswrapper[4802]: I1125 17:10:58.374607 4802 scope.go:117] "RemoveContainer" containerID="2a8bdd0e31e006cf2fc8217352e56a19ec1c338a7e61e7aa598c1a395f5d0822" Nov 25 17:10:58 crc kubenswrapper[4802]: I1125 17:10:58.391608 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7p7fp"] Nov 25 17:10:58 crc kubenswrapper[4802]: I1125 17:10:58.396727 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7p7fp"] Nov 25 17:10:58 crc kubenswrapper[4802]: I1125 17:10:58.407340 4802 scope.go:117] "RemoveContainer" containerID="4b79f5e4296394aaafc2c8f4680acce04662fa63956036fe7e0c6834bfdeb54d" Nov 25 17:10:58 crc kubenswrapper[4802]: I1125 17:10:58.445092 4802 scope.go:117] "RemoveContainer" containerID="30254af7c113bab1a2f7709ddb5cae6f1c236242e67aa88d4ea6c0532141dde0" Nov 25 17:10:58 crc kubenswrapper[4802]: E1125 17:10:58.445614 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30254af7c113bab1a2f7709ddb5cae6f1c236242e67aa88d4ea6c0532141dde0\": container with ID starting with 30254af7c113bab1a2f7709ddb5cae6f1c236242e67aa88d4ea6c0532141dde0 not found: ID does not exist" containerID="30254af7c113bab1a2f7709ddb5cae6f1c236242e67aa88d4ea6c0532141dde0" Nov 25 17:10:58 crc kubenswrapper[4802]: I1125 17:10:58.445667 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30254af7c113bab1a2f7709ddb5cae6f1c236242e67aa88d4ea6c0532141dde0"} err="failed to get container status \"30254af7c113bab1a2f7709ddb5cae6f1c236242e67aa88d4ea6c0532141dde0\": rpc error: code = NotFound desc = could not find container \"30254af7c113bab1a2f7709ddb5cae6f1c236242e67aa88d4ea6c0532141dde0\": container with ID starting with 30254af7c113bab1a2f7709ddb5cae6f1c236242e67aa88d4ea6c0532141dde0 not found: ID does not exist" Nov 25 17:10:58 crc kubenswrapper[4802]: I1125 17:10:58.445700 4802 scope.go:117] "RemoveContainer" containerID="2a8bdd0e31e006cf2fc8217352e56a19ec1c338a7e61e7aa598c1a395f5d0822" Nov 25 17:10:58 crc kubenswrapper[4802]: E1125 17:10:58.445993 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a8bdd0e31e006cf2fc8217352e56a19ec1c338a7e61e7aa598c1a395f5d0822\": container with ID starting with 2a8bdd0e31e006cf2fc8217352e56a19ec1c338a7e61e7aa598c1a395f5d0822 not found: ID does not exist" containerID="2a8bdd0e31e006cf2fc8217352e56a19ec1c338a7e61e7aa598c1a395f5d0822" Nov 25 17:10:58 crc kubenswrapper[4802]: I1125 17:10:58.446022 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a8bdd0e31e006cf2fc8217352e56a19ec1c338a7e61e7aa598c1a395f5d0822"} err="failed to get container status \"2a8bdd0e31e006cf2fc8217352e56a19ec1c338a7e61e7aa598c1a395f5d0822\": rpc error: code = NotFound desc = could not find container \"2a8bdd0e31e006cf2fc8217352e56a19ec1c338a7e61e7aa598c1a395f5d0822\": container with ID starting with 2a8bdd0e31e006cf2fc8217352e56a19ec1c338a7e61e7aa598c1a395f5d0822 not found: ID does not exist" Nov 25 17:10:58 crc kubenswrapper[4802]: I1125 17:10:58.446044 4802 scope.go:117] "RemoveContainer" containerID="4b79f5e4296394aaafc2c8f4680acce04662fa63956036fe7e0c6834bfdeb54d" Nov 25 17:10:58 crc kubenswrapper[4802]: E1125 17:10:58.446260 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b79f5e4296394aaafc2c8f4680acce04662fa63956036fe7e0c6834bfdeb54d\": container with ID starting with 4b79f5e4296394aaafc2c8f4680acce04662fa63956036fe7e0c6834bfdeb54d not found: ID does not exist" containerID="4b79f5e4296394aaafc2c8f4680acce04662fa63956036fe7e0c6834bfdeb54d" Nov 25 17:10:58 crc kubenswrapper[4802]: I1125 17:10:58.446288 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b79f5e4296394aaafc2c8f4680acce04662fa63956036fe7e0c6834bfdeb54d"} err="failed to get container status \"4b79f5e4296394aaafc2c8f4680acce04662fa63956036fe7e0c6834bfdeb54d\": rpc error: code = NotFound desc = could not find container \"4b79f5e4296394aaafc2c8f4680acce04662fa63956036fe7e0c6834bfdeb54d\": container with ID starting with 4b79f5e4296394aaafc2c8f4680acce04662fa63956036fe7e0c6834bfdeb54d not found: ID does not exist" Nov 25 17:10:59 crc kubenswrapper[4802]: I1125 17:10:59.516850 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5018a6f-306c-4866-b72f-1810b53c92ba" path="/var/lib/kubelet/pods/a5018a6f-306c-4866-b72f-1810b53c92ba/volumes" Nov 25 17:11:11 crc kubenswrapper[4802]: I1125 17:11:11.064934 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w4lck"] Nov 25 17:11:11 crc kubenswrapper[4802]: E1125 17:11:11.065845 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5018a6f-306c-4866-b72f-1810b53c92ba" containerName="registry-server" Nov 25 17:11:11 crc kubenswrapper[4802]: I1125 17:11:11.065865 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5018a6f-306c-4866-b72f-1810b53c92ba" containerName="registry-server" Nov 25 17:11:11 crc kubenswrapper[4802]: E1125 17:11:11.065892 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5018a6f-306c-4866-b72f-1810b53c92ba" containerName="extract-content" Nov 25 17:11:11 crc kubenswrapper[4802]: I1125 17:11:11.065900 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5018a6f-306c-4866-b72f-1810b53c92ba" containerName="extract-content" Nov 25 17:11:11 crc kubenswrapper[4802]: E1125 17:11:11.065939 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5018a6f-306c-4866-b72f-1810b53c92ba" containerName="extract-utilities" Nov 25 17:11:11 crc kubenswrapper[4802]: I1125 17:11:11.065949 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5018a6f-306c-4866-b72f-1810b53c92ba" containerName="extract-utilities" Nov 25 17:11:11 crc kubenswrapper[4802]: I1125 17:11:11.066078 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5018a6f-306c-4866-b72f-1810b53c92ba" containerName="registry-server" Nov 25 17:11:11 crc kubenswrapper[4802]: I1125 17:11:11.067174 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w4lck" Nov 25 17:11:11 crc kubenswrapper[4802]: I1125 17:11:11.080805 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w4lck"] Nov 25 17:11:11 crc kubenswrapper[4802]: I1125 17:11:11.236266 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb-catalog-content\") pod \"redhat-operators-w4lck\" (UID: \"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb\") " pod="openshift-marketplace/redhat-operators-w4lck" Nov 25 17:11:11 crc kubenswrapper[4802]: I1125 17:11:11.236308 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb-utilities\") pod \"redhat-operators-w4lck\" (UID: \"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb\") " pod="openshift-marketplace/redhat-operators-w4lck" Nov 25 17:11:11 crc kubenswrapper[4802]: I1125 17:11:11.236355 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7z5q\" (UniqueName: \"kubernetes.io/projected/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb-kube-api-access-p7z5q\") pod \"redhat-operators-w4lck\" (UID: \"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb\") " pod="openshift-marketplace/redhat-operators-w4lck" Nov 25 17:11:11 crc kubenswrapper[4802]: I1125 17:11:11.337788 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb-catalog-content\") pod \"redhat-operators-w4lck\" (UID: \"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb\") " pod="openshift-marketplace/redhat-operators-w4lck" Nov 25 17:11:11 crc kubenswrapper[4802]: I1125 17:11:11.337850 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb-utilities\") pod \"redhat-operators-w4lck\" (UID: \"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb\") " pod="openshift-marketplace/redhat-operators-w4lck" Nov 25 17:11:11 crc kubenswrapper[4802]: I1125 17:11:11.337925 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7z5q\" (UniqueName: \"kubernetes.io/projected/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb-kube-api-access-p7z5q\") pod \"redhat-operators-w4lck\" (UID: \"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb\") " pod="openshift-marketplace/redhat-operators-w4lck" Nov 25 17:11:11 crc kubenswrapper[4802]: I1125 17:11:11.338313 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb-catalog-content\") pod \"redhat-operators-w4lck\" (UID: \"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb\") " pod="openshift-marketplace/redhat-operators-w4lck" Nov 25 17:11:11 crc kubenswrapper[4802]: I1125 17:11:11.338405 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb-utilities\") pod \"redhat-operators-w4lck\" (UID: \"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb\") " pod="openshift-marketplace/redhat-operators-w4lck" Nov 25 17:11:11 crc kubenswrapper[4802]: I1125 17:11:11.369737 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7z5q\" (UniqueName: \"kubernetes.io/projected/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb-kube-api-access-p7z5q\") pod \"redhat-operators-w4lck\" (UID: \"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb\") " pod="openshift-marketplace/redhat-operators-w4lck" Nov 25 17:11:11 crc kubenswrapper[4802]: I1125 17:11:11.394976 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w4lck" Nov 25 17:11:11 crc kubenswrapper[4802]: I1125 17:11:11.827933 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w4lck"] Nov 25 17:11:12 crc kubenswrapper[4802]: I1125 17:11:12.520426 4802 generic.go:334] "Generic (PLEG): container finished" podID="55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb" containerID="a1f93579616c2b78a0fd5745491f19b67f12fdb9933c06075a47897b0fc4c70d" exitCode=0 Nov 25 17:11:12 crc kubenswrapper[4802]: I1125 17:11:12.520469 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w4lck" event={"ID":"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb","Type":"ContainerDied","Data":"a1f93579616c2b78a0fd5745491f19b67f12fdb9933c06075a47897b0fc4c70d"} Nov 25 17:11:12 crc kubenswrapper[4802]: I1125 17:11:12.520494 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w4lck" event={"ID":"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb","Type":"ContainerStarted","Data":"e4d9a2772dce801c13a1cbc3c588b9800d8d17b672ccfa152b9fe11cafae2221"} Nov 25 17:11:14 crc kubenswrapper[4802]: I1125 17:11:14.538187 4802 generic.go:334] "Generic (PLEG): container finished" podID="55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb" containerID="4768bf7dfc216beab7eb90d487ed323c8fe2334694d8c5e31fe5914d93cb737e" exitCode=0 Nov 25 17:11:14 crc kubenswrapper[4802]: I1125 17:11:14.538238 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w4lck" event={"ID":"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb","Type":"ContainerDied","Data":"4768bf7dfc216beab7eb90d487ed323c8fe2334694d8c5e31fe5914d93cb737e"} Nov 25 17:11:16 crc kubenswrapper[4802]: I1125 17:11:16.555992 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w4lck" event={"ID":"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb","Type":"ContainerStarted","Data":"e20df9fad1f2005de072bd2fbbe8d1d21d3c1b51f691ca7194fe77e13b98e5e9"} Nov 25 17:11:16 crc kubenswrapper[4802]: I1125 17:11:16.580879 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w4lck" podStartSLOduration=2.7257646810000002 podStartE2EDuration="5.580861254s" podCreationTimestamp="2025-11-25 17:11:11 +0000 UTC" firstStartedPulling="2025-11-25 17:11:12.522166006 +0000 UTC m=+1455.666513212" lastFinishedPulling="2025-11-25 17:11:15.377262599 +0000 UTC m=+1458.521609785" observedRunningTime="2025-11-25 17:11:16.574915913 +0000 UTC m=+1459.719263099" watchObservedRunningTime="2025-11-25 17:11:16.580861254 +0000 UTC m=+1459.725208440" Nov 25 17:11:18 crc kubenswrapper[4802]: I1125 17:11:18.261706 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cxgc8"] Nov 25 17:11:18 crc kubenswrapper[4802]: I1125 17:11:18.263468 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cxgc8" Nov 25 17:11:18 crc kubenswrapper[4802]: I1125 17:11:18.271086 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cxgc8"] Nov 25 17:11:18 crc kubenswrapper[4802]: I1125 17:11:18.353443 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p7sq\" (UniqueName: \"kubernetes.io/projected/098401de-5466-4330-af6b-0ad15a9cd44b-kube-api-access-5p7sq\") pod \"community-operators-cxgc8\" (UID: \"098401de-5466-4330-af6b-0ad15a9cd44b\") " pod="openshift-marketplace/community-operators-cxgc8" Nov 25 17:11:18 crc kubenswrapper[4802]: I1125 17:11:18.353694 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/098401de-5466-4330-af6b-0ad15a9cd44b-utilities\") pod \"community-operators-cxgc8\" (UID: \"098401de-5466-4330-af6b-0ad15a9cd44b\") " pod="openshift-marketplace/community-operators-cxgc8" Nov 25 17:11:18 crc kubenswrapper[4802]: I1125 17:11:18.353784 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/098401de-5466-4330-af6b-0ad15a9cd44b-catalog-content\") pod \"community-operators-cxgc8\" (UID: \"098401de-5466-4330-af6b-0ad15a9cd44b\") " pod="openshift-marketplace/community-operators-cxgc8" Nov 25 17:11:18 crc kubenswrapper[4802]: I1125 17:11:18.455183 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p7sq\" (UniqueName: \"kubernetes.io/projected/098401de-5466-4330-af6b-0ad15a9cd44b-kube-api-access-5p7sq\") pod \"community-operators-cxgc8\" (UID: \"098401de-5466-4330-af6b-0ad15a9cd44b\") " pod="openshift-marketplace/community-operators-cxgc8" Nov 25 17:11:18 crc kubenswrapper[4802]: I1125 17:11:18.455534 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/098401de-5466-4330-af6b-0ad15a9cd44b-utilities\") pod \"community-operators-cxgc8\" (UID: \"098401de-5466-4330-af6b-0ad15a9cd44b\") " pod="openshift-marketplace/community-operators-cxgc8" Nov 25 17:11:18 crc kubenswrapper[4802]: I1125 17:11:18.455621 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/098401de-5466-4330-af6b-0ad15a9cd44b-catalog-content\") pod \"community-operators-cxgc8\" (UID: \"098401de-5466-4330-af6b-0ad15a9cd44b\") " pod="openshift-marketplace/community-operators-cxgc8" Nov 25 17:11:18 crc kubenswrapper[4802]: I1125 17:11:18.456000 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/098401de-5466-4330-af6b-0ad15a9cd44b-utilities\") pod \"community-operators-cxgc8\" (UID: \"098401de-5466-4330-af6b-0ad15a9cd44b\") " pod="openshift-marketplace/community-operators-cxgc8" Nov 25 17:11:18 crc kubenswrapper[4802]: I1125 17:11:18.456051 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/098401de-5466-4330-af6b-0ad15a9cd44b-catalog-content\") pod \"community-operators-cxgc8\" (UID: \"098401de-5466-4330-af6b-0ad15a9cd44b\") " pod="openshift-marketplace/community-operators-cxgc8" Nov 25 17:11:18 crc kubenswrapper[4802]: I1125 17:11:18.475482 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p7sq\" (UniqueName: \"kubernetes.io/projected/098401de-5466-4330-af6b-0ad15a9cd44b-kube-api-access-5p7sq\") pod \"community-operators-cxgc8\" (UID: \"098401de-5466-4330-af6b-0ad15a9cd44b\") " pod="openshift-marketplace/community-operators-cxgc8" Nov 25 17:11:18 crc kubenswrapper[4802]: I1125 17:11:18.586474 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cxgc8" Nov 25 17:11:19 crc kubenswrapper[4802]: I1125 17:11:19.142866 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cxgc8"] Nov 25 17:11:19 crc kubenswrapper[4802]: W1125 17:11:19.151358 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod098401de_5466_4330_af6b_0ad15a9cd44b.slice/crio-638caa9ec13c920bf7144315b6ee53f71f31852e14d05b0cdf03c9f0d7d62366 WatchSource:0}: Error finding container 638caa9ec13c920bf7144315b6ee53f71f31852e14d05b0cdf03c9f0d7d62366: Status 404 returned error can't find the container with id 638caa9ec13c920bf7144315b6ee53f71f31852e14d05b0cdf03c9f0d7d62366 Nov 25 17:11:19 crc kubenswrapper[4802]: I1125 17:11:19.580636 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cxgc8" event={"ID":"098401de-5466-4330-af6b-0ad15a9cd44b","Type":"ContainerStarted","Data":"17991258f9e6fa9f99bfe1ec1922364e6fe3f2baaf2ebe7e89f0d7f0faedf22c"} Nov 25 17:11:19 crc kubenswrapper[4802]: I1125 17:11:19.580944 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cxgc8" event={"ID":"098401de-5466-4330-af6b-0ad15a9cd44b","Type":"ContainerStarted","Data":"638caa9ec13c920bf7144315b6ee53f71f31852e14d05b0cdf03c9f0d7d62366"} Nov 25 17:11:20 crc kubenswrapper[4802]: I1125 17:11:20.589676 4802 generic.go:334] "Generic (PLEG): container finished" podID="098401de-5466-4330-af6b-0ad15a9cd44b" containerID="17991258f9e6fa9f99bfe1ec1922364e6fe3f2baaf2ebe7e89f0d7f0faedf22c" exitCode=0 Nov 25 17:11:20 crc kubenswrapper[4802]: I1125 17:11:20.589723 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cxgc8" event={"ID":"098401de-5466-4330-af6b-0ad15a9cd44b","Type":"ContainerDied","Data":"17991258f9e6fa9f99bfe1ec1922364e6fe3f2baaf2ebe7e89f0d7f0faedf22c"} Nov 25 17:11:21 crc kubenswrapper[4802]: I1125 17:11:21.395096 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w4lck" Nov 25 17:11:21 crc kubenswrapper[4802]: I1125 17:11:21.395530 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w4lck" Nov 25 17:11:21 crc kubenswrapper[4802]: I1125 17:11:21.439411 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w4lck" Nov 25 17:11:21 crc kubenswrapper[4802]: I1125 17:11:21.601239 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cxgc8" event={"ID":"098401de-5466-4330-af6b-0ad15a9cd44b","Type":"ContainerStarted","Data":"d13ba64dba1e5347a05a4c83aae29213f2b9fc3e833f5fbf983e8e35b72a418b"} Nov 25 17:11:21 crc kubenswrapper[4802]: I1125 17:11:21.659478 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w4lck" Nov 25 17:11:22 crc kubenswrapper[4802]: I1125 17:11:22.612284 4802 generic.go:334] "Generic (PLEG): container finished" podID="098401de-5466-4330-af6b-0ad15a9cd44b" containerID="d13ba64dba1e5347a05a4c83aae29213f2b9fc3e833f5fbf983e8e35b72a418b" exitCode=0 Nov 25 17:11:22 crc kubenswrapper[4802]: I1125 17:11:22.612418 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cxgc8" event={"ID":"098401de-5466-4330-af6b-0ad15a9cd44b","Type":"ContainerDied","Data":"d13ba64dba1e5347a05a4c83aae29213f2b9fc3e833f5fbf983e8e35b72a418b"} Nov 25 17:11:23 crc kubenswrapper[4802]: I1125 17:11:23.621372 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cxgc8" event={"ID":"098401de-5466-4330-af6b-0ad15a9cd44b","Type":"ContainerStarted","Data":"8e16fa3ac568ee3154543458035a6386cbda6879d0b5d957d7581df5221c920b"} Nov 25 17:11:23 crc kubenswrapper[4802]: I1125 17:11:23.655761 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cxgc8" podStartSLOduration=3.246649775 podStartE2EDuration="5.655739635s" podCreationTimestamp="2025-11-25 17:11:18 +0000 UTC" firstStartedPulling="2025-11-25 17:11:20.591104369 +0000 UTC m=+1463.735451555" lastFinishedPulling="2025-11-25 17:11:23.000194199 +0000 UTC m=+1466.144541415" observedRunningTime="2025-11-25 17:11:23.645240811 +0000 UTC m=+1466.789588017" watchObservedRunningTime="2025-11-25 17:11:23.655739635 +0000 UTC m=+1466.800086821" Nov 25 17:11:23 crc kubenswrapper[4802]: I1125 17:11:23.844842 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w4lck"] Nov 25 17:11:23 crc kubenswrapper[4802]: I1125 17:11:23.845144 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w4lck" podUID="55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb" containerName="registry-server" containerID="cri-o://e20df9fad1f2005de072bd2fbbe8d1d21d3c1b51f691ca7194fe77e13b98e5e9" gracePeriod=2 Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.209998 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w4lck" Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.248080 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.248152 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.267442 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb-utilities\") pod \"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb\" (UID: \"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb\") " Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.267537 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb-catalog-content\") pod \"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb\" (UID: \"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb\") " Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.267604 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7z5q\" (UniqueName: \"kubernetes.io/projected/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb-kube-api-access-p7z5q\") pod \"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb\" (UID: \"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb\") " Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.269079 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb-utilities" (OuterVolumeSpecName: "utilities") pod "55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb" (UID: "55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.273247 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb-kube-api-access-p7z5q" (OuterVolumeSpecName: "kube-api-access-p7z5q") pod "55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb" (UID: "55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb"). InnerVolumeSpecName "kube-api-access-p7z5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.369281 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7z5q\" (UniqueName: \"kubernetes.io/projected/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb-kube-api-access-p7z5q\") on node \"crc\" DevicePath \"\"" Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.369310 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.633755 4802 generic.go:334] "Generic (PLEG): container finished" podID="55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb" containerID="e20df9fad1f2005de072bd2fbbe8d1d21d3c1b51f691ca7194fe77e13b98e5e9" exitCode=0 Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.633811 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w4lck" event={"ID":"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb","Type":"ContainerDied","Data":"e20df9fad1f2005de072bd2fbbe8d1d21d3c1b51f691ca7194fe77e13b98e5e9"} Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.633871 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w4lck" Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.633908 4802 scope.go:117] "RemoveContainer" containerID="e20df9fad1f2005de072bd2fbbe8d1d21d3c1b51f691ca7194fe77e13b98e5e9" Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.633889 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w4lck" event={"ID":"55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb","Type":"ContainerDied","Data":"e4d9a2772dce801c13a1cbc3c588b9800d8d17b672ccfa152b9fe11cafae2221"} Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.662283 4802 scope.go:117] "RemoveContainer" containerID="4768bf7dfc216beab7eb90d487ed323c8fe2334694d8c5e31fe5914d93cb737e" Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.692564 4802 scope.go:117] "RemoveContainer" containerID="a1f93579616c2b78a0fd5745491f19b67f12fdb9933c06075a47897b0fc4c70d" Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.728284 4802 scope.go:117] "RemoveContainer" containerID="e20df9fad1f2005de072bd2fbbe8d1d21d3c1b51f691ca7194fe77e13b98e5e9" Nov 25 17:11:24 crc kubenswrapper[4802]: E1125 17:11:24.728902 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e20df9fad1f2005de072bd2fbbe8d1d21d3c1b51f691ca7194fe77e13b98e5e9\": container with ID starting with e20df9fad1f2005de072bd2fbbe8d1d21d3c1b51f691ca7194fe77e13b98e5e9 not found: ID does not exist" containerID="e20df9fad1f2005de072bd2fbbe8d1d21d3c1b51f691ca7194fe77e13b98e5e9" Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.728970 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e20df9fad1f2005de072bd2fbbe8d1d21d3c1b51f691ca7194fe77e13b98e5e9"} err="failed to get container status \"e20df9fad1f2005de072bd2fbbe8d1d21d3c1b51f691ca7194fe77e13b98e5e9\": rpc error: code = NotFound desc = could not find container \"e20df9fad1f2005de072bd2fbbe8d1d21d3c1b51f691ca7194fe77e13b98e5e9\": container with ID starting with e20df9fad1f2005de072bd2fbbe8d1d21d3c1b51f691ca7194fe77e13b98e5e9 not found: ID does not exist" Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.729015 4802 scope.go:117] "RemoveContainer" containerID="4768bf7dfc216beab7eb90d487ed323c8fe2334694d8c5e31fe5914d93cb737e" Nov 25 17:11:24 crc kubenswrapper[4802]: E1125 17:11:24.729568 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4768bf7dfc216beab7eb90d487ed323c8fe2334694d8c5e31fe5914d93cb737e\": container with ID starting with 4768bf7dfc216beab7eb90d487ed323c8fe2334694d8c5e31fe5914d93cb737e not found: ID does not exist" containerID="4768bf7dfc216beab7eb90d487ed323c8fe2334694d8c5e31fe5914d93cb737e" Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.729621 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4768bf7dfc216beab7eb90d487ed323c8fe2334694d8c5e31fe5914d93cb737e"} err="failed to get container status \"4768bf7dfc216beab7eb90d487ed323c8fe2334694d8c5e31fe5914d93cb737e\": rpc error: code = NotFound desc = could not find container \"4768bf7dfc216beab7eb90d487ed323c8fe2334694d8c5e31fe5914d93cb737e\": container with ID starting with 4768bf7dfc216beab7eb90d487ed323c8fe2334694d8c5e31fe5914d93cb737e not found: ID does not exist" Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.729651 4802 scope.go:117] "RemoveContainer" containerID="a1f93579616c2b78a0fd5745491f19b67f12fdb9933c06075a47897b0fc4c70d" Nov 25 17:11:24 crc kubenswrapper[4802]: E1125 17:11:24.730175 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1f93579616c2b78a0fd5745491f19b67f12fdb9933c06075a47897b0fc4c70d\": container with ID starting with a1f93579616c2b78a0fd5745491f19b67f12fdb9933c06075a47897b0fc4c70d not found: ID does not exist" containerID="a1f93579616c2b78a0fd5745491f19b67f12fdb9933c06075a47897b0fc4c70d" Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.730195 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1f93579616c2b78a0fd5745491f19b67f12fdb9933c06075a47897b0fc4c70d"} err="failed to get container status \"a1f93579616c2b78a0fd5745491f19b67f12fdb9933c06075a47897b0fc4c70d\": rpc error: code = NotFound desc = could not find container \"a1f93579616c2b78a0fd5745491f19b67f12fdb9933c06075a47897b0fc4c70d\": container with ID starting with a1f93579616c2b78a0fd5745491f19b67f12fdb9933c06075a47897b0fc4c70d not found: ID does not exist" Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.851948 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb" (UID: "55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.877548 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.967489 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w4lck"] Nov 25 17:11:24 crc kubenswrapper[4802]: I1125 17:11:24.977371 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w4lck"] Nov 25 17:11:25 crc kubenswrapper[4802]: I1125 17:11:25.515154 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb" path="/var/lib/kubelet/pods/55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb/volumes" Nov 25 17:11:28 crc kubenswrapper[4802]: I1125 17:11:28.588996 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cxgc8" Nov 25 17:11:28 crc kubenswrapper[4802]: I1125 17:11:28.589533 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cxgc8" Nov 25 17:11:28 crc kubenswrapper[4802]: I1125 17:11:28.653523 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cxgc8" Nov 25 17:11:28 crc kubenswrapper[4802]: I1125 17:11:28.723466 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cxgc8" Nov 25 17:11:29 crc kubenswrapper[4802]: I1125 17:11:29.848091 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cxgc8"] Nov 25 17:11:30 crc kubenswrapper[4802]: I1125 17:11:30.685535 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cxgc8" podUID="098401de-5466-4330-af6b-0ad15a9cd44b" containerName="registry-server" containerID="cri-o://8e16fa3ac568ee3154543458035a6386cbda6879d0b5d957d7581df5221c920b" gracePeriod=2 Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.613106 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cxgc8" Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.680619 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5p7sq\" (UniqueName: \"kubernetes.io/projected/098401de-5466-4330-af6b-0ad15a9cd44b-kube-api-access-5p7sq\") pod \"098401de-5466-4330-af6b-0ad15a9cd44b\" (UID: \"098401de-5466-4330-af6b-0ad15a9cd44b\") " Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.680683 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/098401de-5466-4330-af6b-0ad15a9cd44b-catalog-content\") pod \"098401de-5466-4330-af6b-0ad15a9cd44b\" (UID: \"098401de-5466-4330-af6b-0ad15a9cd44b\") " Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.680723 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/098401de-5466-4330-af6b-0ad15a9cd44b-utilities\") pod \"098401de-5466-4330-af6b-0ad15a9cd44b\" (UID: \"098401de-5466-4330-af6b-0ad15a9cd44b\") " Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.682034 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/098401de-5466-4330-af6b-0ad15a9cd44b-utilities" (OuterVolumeSpecName: "utilities") pod "098401de-5466-4330-af6b-0ad15a9cd44b" (UID: "098401de-5466-4330-af6b-0ad15a9cd44b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.686596 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/098401de-5466-4330-af6b-0ad15a9cd44b-kube-api-access-5p7sq" (OuterVolumeSpecName: "kube-api-access-5p7sq") pod "098401de-5466-4330-af6b-0ad15a9cd44b" (UID: "098401de-5466-4330-af6b-0ad15a9cd44b"). InnerVolumeSpecName "kube-api-access-5p7sq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.703453 4802 generic.go:334] "Generic (PLEG): container finished" podID="098401de-5466-4330-af6b-0ad15a9cd44b" containerID="8e16fa3ac568ee3154543458035a6386cbda6879d0b5d957d7581df5221c920b" exitCode=0 Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.703497 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cxgc8" event={"ID":"098401de-5466-4330-af6b-0ad15a9cd44b","Type":"ContainerDied","Data":"8e16fa3ac568ee3154543458035a6386cbda6879d0b5d957d7581df5221c920b"} Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.703554 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cxgc8" event={"ID":"098401de-5466-4330-af6b-0ad15a9cd44b","Type":"ContainerDied","Data":"638caa9ec13c920bf7144315b6ee53f71f31852e14d05b0cdf03c9f0d7d62366"} Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.703577 4802 scope.go:117] "RemoveContainer" containerID="8e16fa3ac568ee3154543458035a6386cbda6879d0b5d957d7581df5221c920b" Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.703842 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cxgc8" Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.732618 4802 scope.go:117] "RemoveContainer" containerID="d13ba64dba1e5347a05a4c83aae29213f2b9fc3e833f5fbf983e8e35b72a418b" Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.755024 4802 scope.go:117] "RemoveContainer" containerID="17991258f9e6fa9f99bfe1ec1922364e6fe3f2baaf2ebe7e89f0d7f0faedf22c" Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.756977 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/098401de-5466-4330-af6b-0ad15a9cd44b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "098401de-5466-4330-af6b-0ad15a9cd44b" (UID: "098401de-5466-4330-af6b-0ad15a9cd44b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.782336 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5p7sq\" (UniqueName: \"kubernetes.io/projected/098401de-5466-4330-af6b-0ad15a9cd44b-kube-api-access-5p7sq\") on node \"crc\" DevicePath \"\"" Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.782368 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/098401de-5466-4330-af6b-0ad15a9cd44b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.782377 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/098401de-5466-4330-af6b-0ad15a9cd44b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.784711 4802 scope.go:117] "RemoveContainer" containerID="8e16fa3ac568ee3154543458035a6386cbda6879d0b5d957d7581df5221c920b" Nov 25 17:11:31 crc kubenswrapper[4802]: E1125 17:11:31.785142 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e16fa3ac568ee3154543458035a6386cbda6879d0b5d957d7581df5221c920b\": container with ID starting with 8e16fa3ac568ee3154543458035a6386cbda6879d0b5d957d7581df5221c920b not found: ID does not exist" containerID="8e16fa3ac568ee3154543458035a6386cbda6879d0b5d957d7581df5221c920b" Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.785184 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e16fa3ac568ee3154543458035a6386cbda6879d0b5d957d7581df5221c920b"} err="failed to get container status \"8e16fa3ac568ee3154543458035a6386cbda6879d0b5d957d7581df5221c920b\": rpc error: code = NotFound desc = could not find container \"8e16fa3ac568ee3154543458035a6386cbda6879d0b5d957d7581df5221c920b\": container with ID starting with 8e16fa3ac568ee3154543458035a6386cbda6879d0b5d957d7581df5221c920b not found: ID does not exist" Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.785210 4802 scope.go:117] "RemoveContainer" containerID="d13ba64dba1e5347a05a4c83aae29213f2b9fc3e833f5fbf983e8e35b72a418b" Nov 25 17:11:31 crc kubenswrapper[4802]: E1125 17:11:31.785484 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d13ba64dba1e5347a05a4c83aae29213f2b9fc3e833f5fbf983e8e35b72a418b\": container with ID starting with d13ba64dba1e5347a05a4c83aae29213f2b9fc3e833f5fbf983e8e35b72a418b not found: ID does not exist" containerID="d13ba64dba1e5347a05a4c83aae29213f2b9fc3e833f5fbf983e8e35b72a418b" Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.785547 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d13ba64dba1e5347a05a4c83aae29213f2b9fc3e833f5fbf983e8e35b72a418b"} err="failed to get container status \"d13ba64dba1e5347a05a4c83aae29213f2b9fc3e833f5fbf983e8e35b72a418b\": rpc error: code = NotFound desc = could not find container \"d13ba64dba1e5347a05a4c83aae29213f2b9fc3e833f5fbf983e8e35b72a418b\": container with ID starting with d13ba64dba1e5347a05a4c83aae29213f2b9fc3e833f5fbf983e8e35b72a418b not found: ID does not exist" Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.785570 4802 scope.go:117] "RemoveContainer" containerID="17991258f9e6fa9f99bfe1ec1922364e6fe3f2baaf2ebe7e89f0d7f0faedf22c" Nov 25 17:11:31 crc kubenswrapper[4802]: E1125 17:11:31.785800 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17991258f9e6fa9f99bfe1ec1922364e6fe3f2baaf2ebe7e89f0d7f0faedf22c\": container with ID starting with 17991258f9e6fa9f99bfe1ec1922364e6fe3f2baaf2ebe7e89f0d7f0faedf22c not found: ID does not exist" containerID="17991258f9e6fa9f99bfe1ec1922364e6fe3f2baaf2ebe7e89f0d7f0faedf22c" Nov 25 17:11:31 crc kubenswrapper[4802]: I1125 17:11:31.785825 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17991258f9e6fa9f99bfe1ec1922364e6fe3f2baaf2ebe7e89f0d7f0faedf22c"} err="failed to get container status \"17991258f9e6fa9f99bfe1ec1922364e6fe3f2baaf2ebe7e89f0d7f0faedf22c\": rpc error: code = NotFound desc = could not find container \"17991258f9e6fa9f99bfe1ec1922364e6fe3f2baaf2ebe7e89f0d7f0faedf22c\": container with ID starting with 17991258f9e6fa9f99bfe1ec1922364e6fe3f2baaf2ebe7e89f0d7f0faedf22c not found: ID does not exist" Nov 25 17:11:32 crc kubenswrapper[4802]: I1125 17:11:32.040442 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cxgc8"] Nov 25 17:11:32 crc kubenswrapper[4802]: I1125 17:11:32.045810 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cxgc8"] Nov 25 17:11:33 crc kubenswrapper[4802]: I1125 17:11:33.516666 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="098401de-5466-4330-af6b-0ad15a9cd44b" path="/var/lib/kubelet/pods/098401de-5466-4330-af6b-0ad15a9cd44b/volumes" Nov 25 17:11:54 crc kubenswrapper[4802]: I1125 17:11:54.249189 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 17:11:54 crc kubenswrapper[4802]: I1125 17:11:54.251856 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 17:12:24 crc kubenswrapper[4802]: I1125 17:12:24.249021 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 17:12:24 crc kubenswrapper[4802]: I1125 17:12:24.249511 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 17:12:24 crc kubenswrapper[4802]: I1125 17:12:24.249553 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 17:12:24 crc kubenswrapper[4802]: I1125 17:12:24.250193 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29"} pod="openshift-machine-config-operator/machine-config-daemon-h29wc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 17:12:24 crc kubenswrapper[4802]: I1125 17:12:24.250239 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" containerID="cri-o://28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" gracePeriod=600 Nov 25 17:12:24 crc kubenswrapper[4802]: E1125 17:12:24.419369 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:12:25 crc kubenswrapper[4802]: I1125 17:12:25.147374 4802 generic.go:334] "Generic (PLEG): container finished" podID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" exitCode=0 Nov 25 17:12:25 crc kubenswrapper[4802]: I1125 17:12:25.147425 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerDied","Data":"28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29"} Nov 25 17:12:25 crc kubenswrapper[4802]: I1125 17:12:25.148046 4802 scope.go:117] "RemoveContainer" containerID="ee1883bc854fb1518fe7b71c8f9bba592e9827245176f86eb7e6da8b81dacf06" Nov 25 17:12:25 crc kubenswrapper[4802]: I1125 17:12:25.148856 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:12:25 crc kubenswrapper[4802]: E1125 17:12:25.149208 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:12:39 crc kubenswrapper[4802]: I1125 17:12:39.505567 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:12:39 crc kubenswrapper[4802]: E1125 17:12:39.506391 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:12:49 crc kubenswrapper[4802]: I1125 17:12:49.692672 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 17:12:49 crc kubenswrapper[4802]: I1125 17:12:49.693795 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="a4f07881-4b13-440d-b080-25f1a5df75ac" containerName="glance-log" containerID="cri-o://36f189eb936f1203af5299dde514f6790364ae557f8cb3c191c3c8f93b1b3fa3" gracePeriod=30 Nov 25 17:12:49 crc kubenswrapper[4802]: I1125 17:12:49.693856 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="a4f07881-4b13-440d-b080-25f1a5df75ac" containerName="glance-httpd" containerID="cri-o://7469712d663f353b3939791a9bb5166b011d571df689064f7849ce4c55a0108a" gracePeriod=30 Nov 25 17:12:49 crc kubenswrapper[4802]: I1125 17:12:49.693829 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="a4f07881-4b13-440d-b080-25f1a5df75ac" containerName="glance-api" containerID="cri-o://69b82248fd43aafc0103f6112021d7f1e2e8a14f6332a08cd5e445eee67e29e6" gracePeriod=30 Nov 25 17:12:49 crc kubenswrapper[4802]: I1125 17:12:49.700464 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 17:12:49 crc kubenswrapper[4802]: I1125 17:12:49.700857 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="723777a0-7ac9-42ee-9d5e-95175c0272f9" containerName="glance-log" containerID="cri-o://5b68c1225cb655d3c617515bc47a9a58a5fc3b1c5e871c98dc38166d1ef8a4e9" gracePeriod=30 Nov 25 17:12:49 crc kubenswrapper[4802]: I1125 17:12:49.700962 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="723777a0-7ac9-42ee-9d5e-95175c0272f9" containerName="glance-api" containerID="cri-o://8604befb4306036f25199f027d8899febaa5dabc3aba554146758c29205052ed" gracePeriod=30 Nov 25 17:12:49 crc kubenswrapper[4802]: I1125 17:12:49.701032 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="723777a0-7ac9-42ee-9d5e-95175c0272f9" containerName="glance-httpd" containerID="cri-o://0898fe2b1471d347c1fc6462e60173cb851ab29582cf64c0fdd4629e353bef48" gracePeriod=30 Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.395533 4802 generic.go:334] "Generic (PLEG): container finished" podID="a4f07881-4b13-440d-b080-25f1a5df75ac" containerID="69b82248fd43aafc0103f6112021d7f1e2e8a14f6332a08cd5e445eee67e29e6" exitCode=0 Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.395869 4802 generic.go:334] "Generic (PLEG): container finished" podID="a4f07881-4b13-440d-b080-25f1a5df75ac" containerID="7469712d663f353b3939791a9bb5166b011d571df689064f7849ce4c55a0108a" exitCode=0 Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.395879 4802 generic.go:334] "Generic (PLEG): container finished" podID="a4f07881-4b13-440d-b080-25f1a5df75ac" containerID="36f189eb936f1203af5299dde514f6790364ae557f8cb3c191c3c8f93b1b3fa3" exitCode=143 Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.395868 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"a4f07881-4b13-440d-b080-25f1a5df75ac","Type":"ContainerDied","Data":"69b82248fd43aafc0103f6112021d7f1e2e8a14f6332a08cd5e445eee67e29e6"} Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.395947 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"a4f07881-4b13-440d-b080-25f1a5df75ac","Type":"ContainerDied","Data":"7469712d663f353b3939791a9bb5166b011d571df689064f7849ce4c55a0108a"} Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.395964 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"a4f07881-4b13-440d-b080-25f1a5df75ac","Type":"ContainerDied","Data":"36f189eb936f1203af5299dde514f6790364ae557f8cb3c191c3c8f93b1b3fa3"} Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.399792 4802 generic.go:334] "Generic (PLEG): container finished" podID="723777a0-7ac9-42ee-9d5e-95175c0272f9" containerID="8604befb4306036f25199f027d8899febaa5dabc3aba554146758c29205052ed" exitCode=0 Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.399832 4802 generic.go:334] "Generic (PLEG): container finished" podID="723777a0-7ac9-42ee-9d5e-95175c0272f9" containerID="0898fe2b1471d347c1fc6462e60173cb851ab29582cf64c0fdd4629e353bef48" exitCode=0 Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.399843 4802 generic.go:334] "Generic (PLEG): container finished" podID="723777a0-7ac9-42ee-9d5e-95175c0272f9" containerID="5b68c1225cb655d3c617515bc47a9a58a5fc3b1c5e871c98dc38166d1ef8a4e9" exitCode=143 Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.399866 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"723777a0-7ac9-42ee-9d5e-95175c0272f9","Type":"ContainerDied","Data":"8604befb4306036f25199f027d8899febaa5dabc3aba554146758c29205052ed"} Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.399895 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"723777a0-7ac9-42ee-9d5e-95175c0272f9","Type":"ContainerDied","Data":"0898fe2b1471d347c1fc6462e60173cb851ab29582cf64c0fdd4629e353bef48"} Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.399908 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"723777a0-7ac9-42ee-9d5e-95175c0272f9","Type":"ContainerDied","Data":"5b68c1225cb655d3c617515bc47a9a58a5fc3b1c5e871c98dc38166d1ef8a4e9"} Nov 25 17:12:50 crc kubenswrapper[4802]: E1125 17:12:50.415790 4802 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4f07881_4b13_440d_b080_25f1a5df75ac.slice/crio-conmon-69b82248fd43aafc0103f6112021d7f1e2e8a14f6332a08cd5e445eee67e29e6.scope\": RecentStats: unable to find data in memory cache]" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.521659 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.531333 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642433 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-lib-modules\") pod \"723777a0-7ac9-42ee-9d5e-95175c0272f9\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642499 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fzlp\" (UniqueName: \"kubernetes.io/projected/a4f07881-4b13-440d-b080-25f1a5df75ac-kube-api-access-5fzlp\") pod \"a4f07881-4b13-440d-b080-25f1a5df75ac\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642552 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-lib-modules\") pod \"a4f07881-4b13-440d-b080-25f1a5df75ac\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642580 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-var-locks-brick\") pod \"a4f07881-4b13-440d-b080-25f1a5df75ac\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642607 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggxt2\" (UniqueName: \"kubernetes.io/projected/723777a0-7ac9-42ee-9d5e-95175c0272f9-kube-api-access-ggxt2\") pod \"723777a0-7ac9-42ee-9d5e-95175c0272f9\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642640 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-run\") pod \"a4f07881-4b13-440d-b080-25f1a5df75ac\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642655 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "a4f07881-4b13-440d-b080-25f1a5df75ac" (UID: "a4f07881-4b13-440d-b080-25f1a5df75ac"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642663 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4f07881-4b13-440d-b080-25f1a5df75ac-scripts\") pod \"a4f07881-4b13-440d-b080-25f1a5df75ac\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642706 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "a4f07881-4b13-440d-b080-25f1a5df75ac" (UID: "a4f07881-4b13-440d-b080-25f1a5df75ac"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642746 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-sys" (OuterVolumeSpecName: "sys") pod "a4f07881-4b13-440d-b080-25f1a5df75ac" (UID: "a4f07881-4b13-440d-b080-25f1a5df75ac"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642724 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-sys\") pod \"a4f07881-4b13-440d-b080-25f1a5df75ac\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642766 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-run" (OuterVolumeSpecName: "run") pod "a4f07881-4b13-440d-b080-25f1a5df75ac" (UID: "a4f07881-4b13-440d-b080-25f1a5df75ac"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642792 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-sys\") pod \"723777a0-7ac9-42ee-9d5e-95175c0272f9\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642823 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"a4f07881-4b13-440d-b080-25f1a5df75ac\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642836 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-sys" (OuterVolumeSpecName: "sys") pod "723777a0-7ac9-42ee-9d5e-95175c0272f9" (UID: "723777a0-7ac9-42ee-9d5e-95175c0272f9"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642846 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-etc-nvme\") pod \"723777a0-7ac9-42ee-9d5e-95175c0272f9\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642879 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-run\") pod \"723777a0-7ac9-42ee-9d5e-95175c0272f9\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642908 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a4f07881-4b13-440d-b080-25f1a5df75ac-httpd-run\") pod \"a4f07881-4b13-440d-b080-25f1a5df75ac\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642944 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/723777a0-7ac9-42ee-9d5e-95175c0272f9-httpd-run\") pod \"723777a0-7ac9-42ee-9d5e-95175c0272f9\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642952 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-run" (OuterVolumeSpecName: "run") pod "723777a0-7ac9-42ee-9d5e-95175c0272f9" (UID: "723777a0-7ac9-42ee-9d5e-95175c0272f9"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.642969 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-var-locks-brick\") pod \"723777a0-7ac9-42ee-9d5e-95175c0272f9\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643011 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-dev\") pod \"a4f07881-4b13-440d-b080-25f1a5df75ac\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643000 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "723777a0-7ac9-42ee-9d5e-95175c0272f9" (UID: "723777a0-7ac9-42ee-9d5e-95175c0272f9"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643043 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/723777a0-7ac9-42ee-9d5e-95175c0272f9-logs\") pod \"723777a0-7ac9-42ee-9d5e-95175c0272f9\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643069 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"723777a0-7ac9-42ee-9d5e-95175c0272f9\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643071 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "723777a0-7ac9-42ee-9d5e-95175c0272f9" (UID: "723777a0-7ac9-42ee-9d5e-95175c0272f9"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643092 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-etc-nvme\") pod \"a4f07881-4b13-440d-b080-25f1a5df75ac\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643108 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-dev" (OuterVolumeSpecName: "dev") pod "a4f07881-4b13-440d-b080-25f1a5df75ac" (UID: "a4f07881-4b13-440d-b080-25f1a5df75ac"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643123 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4f07881-4b13-440d-b080-25f1a5df75ac-logs\") pod \"a4f07881-4b13-440d-b080-25f1a5df75ac\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643264 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4f07881-4b13-440d-b080-25f1a5df75ac-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a4f07881-4b13-440d-b080-25f1a5df75ac" (UID: "a4f07881-4b13-440d-b080-25f1a5df75ac"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643310 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "a4f07881-4b13-440d-b080-25f1a5df75ac" (UID: "a4f07881-4b13-440d-b080-25f1a5df75ac"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643335 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/723777a0-7ac9-42ee-9d5e-95175c0272f9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "723777a0-7ac9-42ee-9d5e-95175c0272f9" (UID: "723777a0-7ac9-42ee-9d5e-95175c0272f9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643344 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-etc-iscsi\") pod \"a4f07881-4b13-440d-b080-25f1a5df75ac\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643382 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/723777a0-7ac9-42ee-9d5e-95175c0272f9-config-data\") pod \"723777a0-7ac9-42ee-9d5e-95175c0272f9\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643369 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "a4f07881-4b13-440d-b080-25f1a5df75ac" (UID: "a4f07881-4b13-440d-b080-25f1a5df75ac"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643409 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-etc-iscsi\") pod \"723777a0-7ac9-42ee-9d5e-95175c0272f9\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643443 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"a4f07881-4b13-440d-b080-25f1a5df75ac\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643486 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-dev\") pod \"723777a0-7ac9-42ee-9d5e-95175c0272f9\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643510 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/723777a0-7ac9-42ee-9d5e-95175c0272f9-logs" (OuterVolumeSpecName: "logs") pod "723777a0-7ac9-42ee-9d5e-95175c0272f9" (UID: "723777a0-7ac9-42ee-9d5e-95175c0272f9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643521 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4f07881-4b13-440d-b080-25f1a5df75ac-logs" (OuterVolumeSpecName: "logs") pod "a4f07881-4b13-440d-b080-25f1a5df75ac" (UID: "a4f07881-4b13-440d-b080-25f1a5df75ac"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643538 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4f07881-4b13-440d-b080-25f1a5df75ac-config-data\") pod \"a4f07881-4b13-440d-b080-25f1a5df75ac\" (UID: \"a4f07881-4b13-440d-b080-25f1a5df75ac\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643570 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-dev" (OuterVolumeSpecName: "dev") pod "723777a0-7ac9-42ee-9d5e-95175c0272f9" (UID: "723777a0-7ac9-42ee-9d5e-95175c0272f9"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643572 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/723777a0-7ac9-42ee-9d5e-95175c0272f9-scripts\") pod \"723777a0-7ac9-42ee-9d5e-95175c0272f9\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.643624 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"723777a0-7ac9-42ee-9d5e-95175c0272f9\" (UID: \"723777a0-7ac9-42ee-9d5e-95175c0272f9\") " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.644022 4802 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-sys\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.644042 4802 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-sys\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.644055 4802 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.644065 4802 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.644079 4802 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a4f07881-4b13-440d-b080-25f1a5df75ac-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.644089 4802 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/723777a0-7ac9-42ee-9d5e-95175c0272f9-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.644100 4802 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.644111 4802 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-dev\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.644143 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/723777a0-7ac9-42ee-9d5e-95175c0272f9-logs\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.644155 4802 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.644165 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4f07881-4b13-440d-b080-25f1a5df75ac-logs\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.644179 4802 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.644189 4802 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-dev\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.644201 4802 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.644212 4802 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.644222 4802 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a4f07881-4b13-440d-b080-25f1a5df75ac-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.644918 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "723777a0-7ac9-42ee-9d5e-95175c0272f9" (UID: "723777a0-7ac9-42ee-9d5e-95175c0272f9"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.645064 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "723777a0-7ac9-42ee-9d5e-95175c0272f9" (UID: "723777a0-7ac9-42ee-9d5e-95175c0272f9"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.648387 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "723777a0-7ac9-42ee-9d5e-95175c0272f9" (UID: "723777a0-7ac9-42ee-9d5e-95175c0272f9"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.648696 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4f07881-4b13-440d-b080-25f1a5df75ac-scripts" (OuterVolumeSpecName: "scripts") pod "a4f07881-4b13-440d-b080-25f1a5df75ac" (UID: "a4f07881-4b13-440d-b080-25f1a5df75ac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.649438 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "723777a0-7ac9-42ee-9d5e-95175c0272f9" (UID: "723777a0-7ac9-42ee-9d5e-95175c0272f9"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.649533 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "a4f07881-4b13-440d-b080-25f1a5df75ac" (UID: "a4f07881-4b13-440d-b080-25f1a5df75ac"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.649535 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/723777a0-7ac9-42ee-9d5e-95175c0272f9-scripts" (OuterVolumeSpecName: "scripts") pod "723777a0-7ac9-42ee-9d5e-95175c0272f9" (UID: "723777a0-7ac9-42ee-9d5e-95175c0272f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.650796 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/723777a0-7ac9-42ee-9d5e-95175c0272f9-kube-api-access-ggxt2" (OuterVolumeSpecName: "kube-api-access-ggxt2") pod "723777a0-7ac9-42ee-9d5e-95175c0272f9" (UID: "723777a0-7ac9-42ee-9d5e-95175c0272f9"). InnerVolumeSpecName "kube-api-access-ggxt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.651956 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4f07881-4b13-440d-b080-25f1a5df75ac-kube-api-access-5fzlp" (OuterVolumeSpecName: "kube-api-access-5fzlp") pod "a4f07881-4b13-440d-b080-25f1a5df75ac" (UID: "a4f07881-4b13-440d-b080-25f1a5df75ac"). InnerVolumeSpecName "kube-api-access-5fzlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.652711 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance-cache") pod "a4f07881-4b13-440d-b080-25f1a5df75ac" (UID: "a4f07881-4b13-440d-b080-25f1a5df75ac"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.726733 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/723777a0-7ac9-42ee-9d5e-95175c0272f9-config-data" (OuterVolumeSpecName: "config-data") pod "723777a0-7ac9-42ee-9d5e-95175c0272f9" (UID: "723777a0-7ac9-42ee-9d5e-95175c0272f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.741554 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4f07881-4b13-440d-b080-25f1a5df75ac-config-data" (OuterVolumeSpecName: "config-data") pod "a4f07881-4b13-440d-b080-25f1a5df75ac" (UID: "a4f07881-4b13-440d-b080-25f1a5df75ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.745691 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.745741 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/723777a0-7ac9-42ee-9d5e-95175c0272f9-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.745759 4802 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.745790 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.745803 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4f07881-4b13-440d-b080-25f1a5df75ac-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.745816 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/723777a0-7ac9-42ee-9d5e-95175c0272f9-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.745834 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.745846 4802 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/723777a0-7ac9-42ee-9d5e-95175c0272f9-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.745860 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fzlp\" (UniqueName: \"kubernetes.io/projected/a4f07881-4b13-440d-b080-25f1a5df75ac-kube-api-access-5fzlp\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.745875 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggxt2\" (UniqueName: \"kubernetes.io/projected/723777a0-7ac9-42ee-9d5e-95175c0272f9-kube-api-access-ggxt2\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.745887 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4f07881-4b13-440d-b080-25f1a5df75ac-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.745906 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.770299 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.774007 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.774769 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.775446 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.847432 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.847495 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.847506 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:50 crc kubenswrapper[4802]: I1125 17:12:50.847517 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:12:51 crc kubenswrapper[4802]: I1125 17:12:51.408712 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"723777a0-7ac9-42ee-9d5e-95175c0272f9","Type":"ContainerDied","Data":"b9a543ebc7d3a78a725ff30ba52f56b6569adbec8da2f593758a4674ce73330f"} Nov 25 17:12:51 crc kubenswrapper[4802]: I1125 17:12:51.408790 4802 scope.go:117] "RemoveContainer" containerID="8604befb4306036f25199f027d8899febaa5dabc3aba554146758c29205052ed" Nov 25 17:12:51 crc kubenswrapper[4802]: I1125 17:12:51.408860 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:12:51 crc kubenswrapper[4802]: I1125 17:12:51.410904 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"a4f07881-4b13-440d-b080-25f1a5df75ac","Type":"ContainerDied","Data":"391c01c5b7b031135ea27c5b222f13a7bb0e6a803dcc805821cbde6f9f8de399"} Nov 25 17:12:51 crc kubenswrapper[4802]: I1125 17:12:51.410984 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 17:12:51 crc kubenswrapper[4802]: I1125 17:12:51.428369 4802 scope.go:117] "RemoveContainer" containerID="0898fe2b1471d347c1fc6462e60173cb851ab29582cf64c0fdd4629e353bef48" Nov 25 17:12:51 crc kubenswrapper[4802]: I1125 17:12:51.447168 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 17:12:51 crc kubenswrapper[4802]: I1125 17:12:51.456638 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 17:12:51 crc kubenswrapper[4802]: I1125 17:12:51.460498 4802 scope.go:117] "RemoveContainer" containerID="5b68c1225cb655d3c617515bc47a9a58a5fc3b1c5e871c98dc38166d1ef8a4e9" Nov 25 17:12:51 crc kubenswrapper[4802]: I1125 17:12:51.460920 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 17:12:51 crc kubenswrapper[4802]: I1125 17:12:51.466156 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 17:12:51 crc kubenswrapper[4802]: I1125 17:12:51.481954 4802 scope.go:117] "RemoveContainer" containerID="69b82248fd43aafc0103f6112021d7f1e2e8a14f6332a08cd5e445eee67e29e6" Nov 25 17:12:51 crc kubenswrapper[4802]: I1125 17:12:51.499988 4802 scope.go:117] "RemoveContainer" containerID="7469712d663f353b3939791a9bb5166b011d571df689064f7849ce4c55a0108a" Nov 25 17:12:51 crc kubenswrapper[4802]: I1125 17:12:51.513252 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="723777a0-7ac9-42ee-9d5e-95175c0272f9" path="/var/lib/kubelet/pods/723777a0-7ac9-42ee-9d5e-95175c0272f9/volumes" Nov 25 17:12:51 crc kubenswrapper[4802]: I1125 17:12:51.514768 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4f07881-4b13-440d-b080-25f1a5df75ac" path="/var/lib/kubelet/pods/a4f07881-4b13-440d-b080-25f1a5df75ac/volumes" Nov 25 17:12:51 crc kubenswrapper[4802]: I1125 17:12:51.519994 4802 scope.go:117] "RemoveContainer" containerID="36f189eb936f1203af5299dde514f6790364ae557f8cb3c191c3c8f93b1b3fa3" Nov 25 17:12:54 crc kubenswrapper[4802]: I1125 17:12:54.504310 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:12:54 crc kubenswrapper[4802]: E1125 17:12:54.505059 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:13:03 crc kubenswrapper[4802]: I1125 17:13:03.774722 4802 scope.go:117] "RemoveContainer" containerID="bcb0d353804d606af76bda79abe8eb0205565231c182c508116e7f9b967a3671" Nov 25 17:13:03 crc kubenswrapper[4802]: I1125 17:13:03.793306 4802 scope.go:117] "RemoveContainer" containerID="d73f32305be41b0056e46f742b1c53aeb8f76f9cb88a65a5cd53e0539f2353a4" Nov 25 17:13:03 crc kubenswrapper[4802]: I1125 17:13:03.822531 4802 scope.go:117] "RemoveContainer" containerID="faf65e25c9f071d1e0ebfb2348654e108cf3fbc584391de072ac29e6b7831a80" Nov 25 17:13:09 crc kubenswrapper[4802]: I1125 17:13:09.505499 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:13:09 crc kubenswrapper[4802]: E1125 17:13:09.508594 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:13:21 crc kubenswrapper[4802]: I1125 17:13:21.505554 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:13:21 crc kubenswrapper[4802]: E1125 17:13:21.506884 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:13:32 crc kubenswrapper[4802]: I1125 17:13:32.505262 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:13:32 crc kubenswrapper[4802]: E1125 17:13:32.506246 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:13:45 crc kubenswrapper[4802]: I1125 17:13:45.055888 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-create-gzvs6"] Nov 25 17:13:45 crc kubenswrapper[4802]: I1125 17:13:45.063673 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-create-gzvs6"] Nov 25 17:13:45 crc kubenswrapper[4802]: I1125 17:13:45.513835 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2004be61-2b02-4bc4-a79e-5beba8cb4a60" path="/var/lib/kubelet/pods/2004be61-2b02-4bc4-a79e-5beba8cb4a60/volumes" Nov 25 17:13:46 crc kubenswrapper[4802]: I1125 17:13:46.031905 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-6222-account-create-update-qzmx5"] Nov 25 17:13:46 crc kubenswrapper[4802]: I1125 17:13:46.039421 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-6222-account-create-update-qzmx5"] Nov 25 17:13:47 crc kubenswrapper[4802]: I1125 17:13:47.509782 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:13:47 crc kubenswrapper[4802]: E1125 17:13:47.510057 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:13:47 crc kubenswrapper[4802]: I1125 17:13:47.514060 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e6a7e14-e3b5-4105-8efb-25f2fd871dd6" path="/var/lib/kubelet/pods/9e6a7e14-e3b5-4105-8efb-25f2fd871dd6/volumes" Nov 25 17:13:58 crc kubenswrapper[4802]: I1125 17:13:58.505226 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:13:58 crc kubenswrapper[4802]: E1125 17:13:58.506089 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:14:03 crc kubenswrapper[4802]: I1125 17:14:03.940285 4802 scope.go:117] "RemoveContainer" containerID="e3c7904150153962589b5788cab2aa5b01c3f28c73938dde9f6e0b9c94461b54" Nov 25 17:14:03 crc kubenswrapper[4802]: I1125 17:14:03.976112 4802 scope.go:117] "RemoveContainer" containerID="07fe37447f8badd570e3180f6c54c100c16950d5cf9975117ee0cac3a59616bf" Nov 25 17:14:03 crc kubenswrapper[4802]: I1125 17:14:03.996856 4802 scope.go:117] "RemoveContainer" containerID="a80bf8be329c35e0a009bd66ada0f49aeb9fd69ec21cbca077fd1c16637f88ba" Nov 25 17:14:04 crc kubenswrapper[4802]: I1125 17:14:04.020854 4802 scope.go:117] "RemoveContainer" containerID="02651fd5d18544c34af0b94b5587fac8323de356e54654c775aec55f7ad7e08c" Nov 25 17:14:04 crc kubenswrapper[4802]: I1125 17:14:04.102607 4802 scope.go:117] "RemoveContainer" containerID="6d230defa21cecb2df57a1df15e4866fc3644483b542b239c2a0dfbb84e2bff3" Nov 25 17:14:04 crc kubenswrapper[4802]: I1125 17:14:04.121300 4802 scope.go:117] "RemoveContainer" containerID="9c3ba8f730ea71ef73f67284f774443dfb7dd6eea6ae7fbd8b2f4c97b17fac29" Nov 25 17:14:09 crc kubenswrapper[4802]: I1125 17:14:09.504604 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:14:09 crc kubenswrapper[4802]: E1125 17:14:09.507189 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:14:11 crc kubenswrapper[4802]: I1125 17:14:11.046135 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-7ghgc"] Nov 25 17:14:11 crc kubenswrapper[4802]: I1125 17:14:11.052396 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-7ghgc"] Nov 25 17:14:11 crc kubenswrapper[4802]: I1125 17:14:11.514891 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75eae88a-5232-4493-9fa1-895961fb3d26" path="/var/lib/kubelet/pods/75eae88a-5232-4493-9fa1-895961fb3d26/volumes" Nov 25 17:14:18 crc kubenswrapper[4802]: I1125 17:14:18.050931 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-7czbg"] Nov 25 17:14:18 crc kubenswrapper[4802]: I1125 17:14:18.058307 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-7czbg"] Nov 25 17:14:19 crc kubenswrapper[4802]: I1125 17:14:19.520452 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9278966-eb40-4107-b492-2012500b1701" path="/var/lib/kubelet/pods/d9278966-eb40-4107-b492-2012500b1701/volumes" Nov 25 17:14:21 crc kubenswrapper[4802]: I1125 17:14:21.505447 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:14:21 crc kubenswrapper[4802]: E1125 17:14:21.505786 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:14:32 crc kubenswrapper[4802]: I1125 17:14:32.504074 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:14:32 crc kubenswrapper[4802]: E1125 17:14:32.504805 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:14:46 crc kubenswrapper[4802]: I1125 17:14:46.504413 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:14:46 crc kubenswrapper[4802]: E1125 17:14:46.505339 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.091956 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 17:14:49 crc kubenswrapper[4802]: E1125 17:14:49.092365 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb" containerName="extract-utilities" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.092384 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb" containerName="extract-utilities" Nov 25 17:14:49 crc kubenswrapper[4802]: E1125 17:14:49.092411 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4f07881-4b13-440d-b080-25f1a5df75ac" containerName="glance-log" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.092425 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4f07881-4b13-440d-b080-25f1a5df75ac" containerName="glance-log" Nov 25 17:14:49 crc kubenswrapper[4802]: E1125 17:14:49.092451 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="098401de-5466-4330-af6b-0ad15a9cd44b" containerName="registry-server" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.092465 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="098401de-5466-4330-af6b-0ad15a9cd44b" containerName="registry-server" Nov 25 17:14:49 crc kubenswrapper[4802]: E1125 17:14:49.092493 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="723777a0-7ac9-42ee-9d5e-95175c0272f9" containerName="glance-log" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.092506 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="723777a0-7ac9-42ee-9d5e-95175c0272f9" containerName="glance-log" Nov 25 17:14:49 crc kubenswrapper[4802]: E1125 17:14:49.092529 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="723777a0-7ac9-42ee-9d5e-95175c0272f9" containerName="glance-httpd" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.092541 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="723777a0-7ac9-42ee-9d5e-95175c0272f9" containerName="glance-httpd" Nov 25 17:14:49 crc kubenswrapper[4802]: E1125 17:14:49.092569 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4f07881-4b13-440d-b080-25f1a5df75ac" containerName="glance-httpd" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.092581 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4f07881-4b13-440d-b080-25f1a5df75ac" containerName="glance-httpd" Nov 25 17:14:49 crc kubenswrapper[4802]: E1125 17:14:49.092615 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="098401de-5466-4330-af6b-0ad15a9cd44b" containerName="extract-content" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.092626 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="098401de-5466-4330-af6b-0ad15a9cd44b" containerName="extract-content" Nov 25 17:14:49 crc kubenswrapper[4802]: E1125 17:14:49.092647 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4f07881-4b13-440d-b080-25f1a5df75ac" containerName="glance-api" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.092658 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4f07881-4b13-440d-b080-25f1a5df75ac" containerName="glance-api" Nov 25 17:14:49 crc kubenswrapper[4802]: E1125 17:14:49.092682 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="098401de-5466-4330-af6b-0ad15a9cd44b" containerName="extract-utilities" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.092695 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="098401de-5466-4330-af6b-0ad15a9cd44b" containerName="extract-utilities" Nov 25 17:14:49 crc kubenswrapper[4802]: E1125 17:14:49.092714 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="723777a0-7ac9-42ee-9d5e-95175c0272f9" containerName="glance-api" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.092726 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="723777a0-7ac9-42ee-9d5e-95175c0272f9" containerName="glance-api" Nov 25 17:14:49 crc kubenswrapper[4802]: E1125 17:14:49.092751 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb" containerName="extract-content" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.092763 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb" containerName="extract-content" Nov 25 17:14:49 crc kubenswrapper[4802]: E1125 17:14:49.092784 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb" containerName="registry-server" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.092796 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb" containerName="registry-server" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.092999 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="098401de-5466-4330-af6b-0ad15a9cd44b" containerName="registry-server" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.093025 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4f07881-4b13-440d-b080-25f1a5df75ac" containerName="glance-api" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.093043 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="55bc4f36-e7dd-4cfd-b6ad-9a20ce9988bb" containerName="registry-server" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.093058 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="723777a0-7ac9-42ee-9d5e-95175c0272f9" containerName="glance-log" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.093080 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="723777a0-7ac9-42ee-9d5e-95175c0272f9" containerName="glance-api" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.093102 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="723777a0-7ac9-42ee-9d5e-95175c0272f9" containerName="glance-httpd" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.093142 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4f07881-4b13-440d-b080-25f1a5df75ac" containerName="glance-httpd" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.093160 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4f07881-4b13-440d-b080-25f1a5df75ac" containerName="glance-log" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.093872 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.101723 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.102037 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.131852 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.195570 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5deabbde-4255-4c46-af2e-22c4e65ecda2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5deabbde-4255-4c46-af2e-22c4e65ecda2\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.195642 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5deabbde-4255-4c46-af2e-22c4e65ecda2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5deabbde-4255-4c46-af2e-22c4e65ecda2\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.297316 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5deabbde-4255-4c46-af2e-22c4e65ecda2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5deabbde-4255-4c46-af2e-22c4e65ecda2\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.297387 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5deabbde-4255-4c46-af2e-22c4e65ecda2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5deabbde-4255-4c46-af2e-22c4e65ecda2\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.297467 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5deabbde-4255-4c46-af2e-22c4e65ecda2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5deabbde-4255-4c46-af2e-22c4e65ecda2\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.319782 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5deabbde-4255-4c46-af2e-22c4e65ecda2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5deabbde-4255-4c46-af2e-22c4e65ecda2\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.417160 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 17:14:49 crc kubenswrapper[4802]: I1125 17:14:49.861899 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 17:14:50 crc kubenswrapper[4802]: I1125 17:14:50.452557 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"5deabbde-4255-4c46-af2e-22c4e65ecda2","Type":"ContainerStarted","Data":"03e684067eb4bfcaf1c50bc57a03f9b7818d0ca39b2685ef12dd4f8a541d52c8"} Nov 25 17:14:50 crc kubenswrapper[4802]: I1125 17:14:50.452921 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"5deabbde-4255-4c46-af2e-22c4e65ecda2","Type":"ContainerStarted","Data":"666869848edb9ae25c020f7434d3495ea4ba30287618272707900305418d48dd"} Nov 25 17:14:50 crc kubenswrapper[4802]: I1125 17:14:50.467815 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=1.467795675 podStartE2EDuration="1.467795675s" podCreationTimestamp="2025-11-25 17:14:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:14:50.465151473 +0000 UTC m=+1673.609498659" watchObservedRunningTime="2025-11-25 17:14:50.467795675 +0000 UTC m=+1673.612142881" Nov 25 17:14:51 crc kubenswrapper[4802]: I1125 17:14:51.462838 4802 generic.go:334] "Generic (PLEG): container finished" podID="5deabbde-4255-4c46-af2e-22c4e65ecda2" containerID="03e684067eb4bfcaf1c50bc57a03f9b7818d0ca39b2685ef12dd4f8a541d52c8" exitCode=0 Nov 25 17:14:51 crc kubenswrapper[4802]: I1125 17:14:51.462901 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"5deabbde-4255-4c46-af2e-22c4e65ecda2","Type":"ContainerDied","Data":"03e684067eb4bfcaf1c50bc57a03f9b7818d0ca39b2685ef12dd4f8a541d52c8"} Nov 25 17:14:52 crc kubenswrapper[4802]: I1125 17:14:52.786281 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 17:14:52 crc kubenswrapper[4802]: I1125 17:14:52.952252 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5deabbde-4255-4c46-af2e-22c4e65ecda2-kubelet-dir\") pod \"5deabbde-4255-4c46-af2e-22c4e65ecda2\" (UID: \"5deabbde-4255-4c46-af2e-22c4e65ecda2\") " Nov 25 17:14:52 crc kubenswrapper[4802]: I1125 17:14:52.952744 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5deabbde-4255-4c46-af2e-22c4e65ecda2-kube-api-access\") pod \"5deabbde-4255-4c46-af2e-22c4e65ecda2\" (UID: \"5deabbde-4255-4c46-af2e-22c4e65ecda2\") " Nov 25 17:14:52 crc kubenswrapper[4802]: I1125 17:14:52.952390 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5deabbde-4255-4c46-af2e-22c4e65ecda2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5deabbde-4255-4c46-af2e-22c4e65ecda2" (UID: "5deabbde-4255-4c46-af2e-22c4e65ecda2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:14:52 crc kubenswrapper[4802]: I1125 17:14:52.953451 4802 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5deabbde-4255-4c46-af2e-22c4e65ecda2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 17:14:52 crc kubenswrapper[4802]: I1125 17:14:52.957468 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5deabbde-4255-4c46-af2e-22c4e65ecda2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5deabbde-4255-4c46-af2e-22c4e65ecda2" (UID: "5deabbde-4255-4c46-af2e-22c4e65ecda2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:14:53 crc kubenswrapper[4802]: I1125 17:14:53.055067 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5deabbde-4255-4c46-af2e-22c4e65ecda2-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 17:14:53 crc kubenswrapper[4802]: I1125 17:14:53.485165 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"5deabbde-4255-4c46-af2e-22c4e65ecda2","Type":"ContainerDied","Data":"666869848edb9ae25c020f7434d3495ea4ba30287618272707900305418d48dd"} Nov 25 17:14:53 crc kubenswrapper[4802]: I1125 17:14:53.485201 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="666869848edb9ae25c020f7434d3495ea4ba30287618272707900305418d48dd" Nov 25 17:14:53 crc kubenswrapper[4802]: I1125 17:14:53.485255 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 17:14:53 crc kubenswrapper[4802]: I1125 17:14:53.888106 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 17:14:53 crc kubenswrapper[4802]: E1125 17:14:53.888453 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5deabbde-4255-4c46-af2e-22c4e65ecda2" containerName="pruner" Nov 25 17:14:53 crc kubenswrapper[4802]: I1125 17:14:53.888466 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="5deabbde-4255-4c46-af2e-22c4e65ecda2" containerName="pruner" Nov 25 17:14:53 crc kubenswrapper[4802]: I1125 17:14:53.888625 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="5deabbde-4255-4c46-af2e-22c4e65ecda2" containerName="pruner" Nov 25 17:14:53 crc kubenswrapper[4802]: I1125 17:14:53.889174 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 17:14:53 crc kubenswrapper[4802]: I1125 17:14:53.893180 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 17:14:53 crc kubenswrapper[4802]: I1125 17:14:53.893451 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 17:14:53 crc kubenswrapper[4802]: I1125 17:14:53.900760 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 17:14:54 crc kubenswrapper[4802]: I1125 17:14:54.069009 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7a7f65e4-cae3-49e6-8f88-df9a7e115260-var-lock\") pod \"installer-9-crc\" (UID: \"7a7f65e4-cae3-49e6-8f88-df9a7e115260\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 17:14:54 crc kubenswrapper[4802]: I1125 17:14:54.069352 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7a7f65e4-cae3-49e6-8f88-df9a7e115260-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7a7f65e4-cae3-49e6-8f88-df9a7e115260\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 17:14:54 crc kubenswrapper[4802]: I1125 17:14:54.069535 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a7f65e4-cae3-49e6-8f88-df9a7e115260-kube-api-access\") pod \"installer-9-crc\" (UID: \"7a7f65e4-cae3-49e6-8f88-df9a7e115260\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 17:14:54 crc kubenswrapper[4802]: I1125 17:14:54.171175 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a7f65e4-cae3-49e6-8f88-df9a7e115260-kube-api-access\") pod \"installer-9-crc\" (UID: \"7a7f65e4-cae3-49e6-8f88-df9a7e115260\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 17:14:54 crc kubenswrapper[4802]: I1125 17:14:54.171258 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7a7f65e4-cae3-49e6-8f88-df9a7e115260-var-lock\") pod \"installer-9-crc\" (UID: \"7a7f65e4-cae3-49e6-8f88-df9a7e115260\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 17:14:54 crc kubenswrapper[4802]: I1125 17:14:54.171308 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7a7f65e4-cae3-49e6-8f88-df9a7e115260-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7a7f65e4-cae3-49e6-8f88-df9a7e115260\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 17:14:54 crc kubenswrapper[4802]: I1125 17:14:54.171414 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7a7f65e4-cae3-49e6-8f88-df9a7e115260-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7a7f65e4-cae3-49e6-8f88-df9a7e115260\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 17:14:54 crc kubenswrapper[4802]: I1125 17:14:54.171460 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7a7f65e4-cae3-49e6-8f88-df9a7e115260-var-lock\") pod \"installer-9-crc\" (UID: \"7a7f65e4-cae3-49e6-8f88-df9a7e115260\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 17:14:54 crc kubenswrapper[4802]: I1125 17:14:54.193623 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a7f65e4-cae3-49e6-8f88-df9a7e115260-kube-api-access\") pod \"installer-9-crc\" (UID: \"7a7f65e4-cae3-49e6-8f88-df9a7e115260\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 17:14:54 crc kubenswrapper[4802]: I1125 17:14:54.218788 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 17:14:54 crc kubenswrapper[4802]: I1125 17:14:54.648352 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 17:14:55 crc kubenswrapper[4802]: I1125 17:14:55.502973 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7a7f65e4-cae3-49e6-8f88-df9a7e115260","Type":"ContainerStarted","Data":"ce87fcfa6d3a6ade986c6bc2252d8dbc050b6b4ceb82c02d31b1612e9df4935c"} Nov 25 17:14:55 crc kubenswrapper[4802]: I1125 17:14:55.503377 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7a7f65e4-cae3-49e6-8f88-df9a7e115260","Type":"ContainerStarted","Data":"9fa31319a54e88d94c034ee5c4478d400fa958b65506fffc31ebb8dd3dbba150"} Nov 25 17:14:55 crc kubenswrapper[4802]: I1125 17:14:55.536868 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.536849008 podStartE2EDuration="2.536849008s" podCreationTimestamp="2025-11-25 17:14:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:14:55.531979237 +0000 UTC m=+1678.676326463" watchObservedRunningTime="2025-11-25 17:14:55.536849008 +0000 UTC m=+1678.681196204" Nov 25 17:14:59 crc kubenswrapper[4802]: I1125 17:14:59.505085 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:14:59 crc kubenswrapper[4802]: E1125 17:14:59.505925 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:15:00 crc kubenswrapper[4802]: I1125 17:15:00.135205 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5"] Nov 25 17:15:00 crc kubenswrapper[4802]: I1125 17:15:00.136324 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5" Nov 25 17:15:00 crc kubenswrapper[4802]: I1125 17:15:00.138139 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 17:15:00 crc kubenswrapper[4802]: I1125 17:15:00.138374 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 17:15:00 crc kubenswrapper[4802]: I1125 17:15:00.146053 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5"] Nov 25 17:15:00 crc kubenswrapper[4802]: I1125 17:15:00.264619 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c5ad4db-e797-4c07-885b-df111a70e0d0-config-volume\") pod \"collect-profiles-29401515-j79v5\" (UID: \"8c5ad4db-e797-4c07-885b-df111a70e0d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5" Nov 25 17:15:00 crc kubenswrapper[4802]: I1125 17:15:00.264679 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lb8l\" (UniqueName: \"kubernetes.io/projected/8c5ad4db-e797-4c07-885b-df111a70e0d0-kube-api-access-7lb8l\") pod \"collect-profiles-29401515-j79v5\" (UID: \"8c5ad4db-e797-4c07-885b-df111a70e0d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5" Nov 25 17:15:00 crc kubenswrapper[4802]: I1125 17:15:00.264713 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c5ad4db-e797-4c07-885b-df111a70e0d0-secret-volume\") pod \"collect-profiles-29401515-j79v5\" (UID: \"8c5ad4db-e797-4c07-885b-df111a70e0d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5" Nov 25 17:15:00 crc kubenswrapper[4802]: I1125 17:15:00.365888 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c5ad4db-e797-4c07-885b-df111a70e0d0-secret-volume\") pod \"collect-profiles-29401515-j79v5\" (UID: \"8c5ad4db-e797-4c07-885b-df111a70e0d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5" Nov 25 17:15:00 crc kubenswrapper[4802]: I1125 17:15:00.366004 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c5ad4db-e797-4c07-885b-df111a70e0d0-config-volume\") pod \"collect-profiles-29401515-j79v5\" (UID: \"8c5ad4db-e797-4c07-885b-df111a70e0d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5" Nov 25 17:15:00 crc kubenswrapper[4802]: I1125 17:15:00.366028 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lb8l\" (UniqueName: \"kubernetes.io/projected/8c5ad4db-e797-4c07-885b-df111a70e0d0-kube-api-access-7lb8l\") pod \"collect-profiles-29401515-j79v5\" (UID: \"8c5ad4db-e797-4c07-885b-df111a70e0d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5" Nov 25 17:15:00 crc kubenswrapper[4802]: I1125 17:15:00.367597 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c5ad4db-e797-4c07-885b-df111a70e0d0-config-volume\") pod \"collect-profiles-29401515-j79v5\" (UID: \"8c5ad4db-e797-4c07-885b-df111a70e0d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5" Nov 25 17:15:00 crc kubenswrapper[4802]: I1125 17:15:00.372475 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c5ad4db-e797-4c07-885b-df111a70e0d0-secret-volume\") pod \"collect-profiles-29401515-j79v5\" (UID: \"8c5ad4db-e797-4c07-885b-df111a70e0d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5" Nov 25 17:15:00 crc kubenswrapper[4802]: I1125 17:15:00.382916 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lb8l\" (UniqueName: \"kubernetes.io/projected/8c5ad4db-e797-4c07-885b-df111a70e0d0-kube-api-access-7lb8l\") pod \"collect-profiles-29401515-j79v5\" (UID: \"8c5ad4db-e797-4c07-885b-df111a70e0d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5" Nov 25 17:15:00 crc kubenswrapper[4802]: I1125 17:15:00.454423 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5" Nov 25 17:15:00 crc kubenswrapper[4802]: I1125 17:15:00.877675 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5"] Nov 25 17:15:01 crc kubenswrapper[4802]: I1125 17:15:01.546903 4802 generic.go:334] "Generic (PLEG): container finished" podID="8c5ad4db-e797-4c07-885b-df111a70e0d0" containerID="9666e1dce13446905dc784866bd37ad1a3b4fd3e6ae845d56a9570853f645670" exitCode=0 Nov 25 17:15:01 crc kubenswrapper[4802]: I1125 17:15:01.546957 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5" event={"ID":"8c5ad4db-e797-4c07-885b-df111a70e0d0","Type":"ContainerDied","Data":"9666e1dce13446905dc784866bd37ad1a3b4fd3e6ae845d56a9570853f645670"} Nov 25 17:15:01 crc kubenswrapper[4802]: I1125 17:15:01.546984 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5" event={"ID":"8c5ad4db-e797-4c07-885b-df111a70e0d0","Type":"ContainerStarted","Data":"9baaa4c70c76b53bf5ebd4894699cafc5e9447acf921de92cb3e44782cfe7e3c"} Nov 25 17:15:02 crc kubenswrapper[4802]: I1125 17:15:02.827759 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5" Nov 25 17:15:02 crc kubenswrapper[4802]: I1125 17:15:02.902968 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lb8l\" (UniqueName: \"kubernetes.io/projected/8c5ad4db-e797-4c07-885b-df111a70e0d0-kube-api-access-7lb8l\") pod \"8c5ad4db-e797-4c07-885b-df111a70e0d0\" (UID: \"8c5ad4db-e797-4c07-885b-df111a70e0d0\") " Nov 25 17:15:02 crc kubenswrapper[4802]: I1125 17:15:02.903042 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c5ad4db-e797-4c07-885b-df111a70e0d0-config-volume\") pod \"8c5ad4db-e797-4c07-885b-df111a70e0d0\" (UID: \"8c5ad4db-e797-4c07-885b-df111a70e0d0\") " Nov 25 17:15:02 crc kubenswrapper[4802]: I1125 17:15:02.903119 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c5ad4db-e797-4c07-885b-df111a70e0d0-secret-volume\") pod \"8c5ad4db-e797-4c07-885b-df111a70e0d0\" (UID: \"8c5ad4db-e797-4c07-885b-df111a70e0d0\") " Nov 25 17:15:02 crc kubenswrapper[4802]: I1125 17:15:02.903801 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c5ad4db-e797-4c07-885b-df111a70e0d0-config-volume" (OuterVolumeSpecName: "config-volume") pod "8c5ad4db-e797-4c07-885b-df111a70e0d0" (UID: "8c5ad4db-e797-4c07-885b-df111a70e0d0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:15:02 crc kubenswrapper[4802]: I1125 17:15:02.909931 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c5ad4db-e797-4c07-885b-df111a70e0d0-kube-api-access-7lb8l" (OuterVolumeSpecName: "kube-api-access-7lb8l") pod "8c5ad4db-e797-4c07-885b-df111a70e0d0" (UID: "8c5ad4db-e797-4c07-885b-df111a70e0d0"). InnerVolumeSpecName "kube-api-access-7lb8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:15:02 crc kubenswrapper[4802]: I1125 17:15:02.910443 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c5ad4db-e797-4c07-885b-df111a70e0d0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8c5ad4db-e797-4c07-885b-df111a70e0d0" (UID: "8c5ad4db-e797-4c07-885b-df111a70e0d0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:15:03 crc kubenswrapper[4802]: I1125 17:15:03.004117 4802 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c5ad4db-e797-4c07-885b-df111a70e0d0-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 17:15:03 crc kubenswrapper[4802]: I1125 17:15:03.004197 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lb8l\" (UniqueName: \"kubernetes.io/projected/8c5ad4db-e797-4c07-885b-df111a70e0d0-kube-api-access-7lb8l\") on node \"crc\" DevicePath \"\"" Nov 25 17:15:03 crc kubenswrapper[4802]: I1125 17:15:03.004209 4802 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c5ad4db-e797-4c07-885b-df111a70e0d0-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 17:15:03 crc kubenswrapper[4802]: I1125 17:15:03.572472 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5" event={"ID":"8c5ad4db-e797-4c07-885b-df111a70e0d0","Type":"ContainerDied","Data":"9baaa4c70c76b53bf5ebd4894699cafc5e9447acf921de92cb3e44782cfe7e3c"} Nov 25 17:15:03 crc kubenswrapper[4802]: I1125 17:15:03.572525 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9baaa4c70c76b53bf5ebd4894699cafc5e9447acf921de92cb3e44782cfe7e3c" Nov 25 17:15:03 crc kubenswrapper[4802]: I1125 17:15:03.572599 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401515-j79v5" Nov 25 17:15:04 crc kubenswrapper[4802]: I1125 17:15:04.241636 4802 scope.go:117] "RemoveContainer" containerID="162bdc1af7a8245196ce544a3deff91163dd4f10dd7669e0b6a1ab8a42921fad" Nov 25 17:15:04 crc kubenswrapper[4802]: I1125 17:15:04.286262 4802 scope.go:117] "RemoveContainer" containerID="5c0e9c2c113faea7b3e3112544b28b3ea5a43e4b6e94989a8c183a64a79b22d1" Nov 25 17:15:04 crc kubenswrapper[4802]: I1125 17:15:04.307305 4802 scope.go:117] "RemoveContainer" containerID="8cec2010a2a8ba7ea562b05c08b0877d5544ac04385e73e610ffa50e4bc88dbe" Nov 25 17:15:12 crc kubenswrapper[4802]: I1125 17:15:12.505009 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:15:12 crc kubenswrapper[4802]: E1125 17:15:12.505766 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:15:26 crc kubenswrapper[4802]: I1125 17:15:26.504233 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:15:26 crc kubenswrapper[4802]: E1125 17:15:26.505011 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.915193 4802 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 17:15:32 crc kubenswrapper[4802]: E1125 17:15:32.916087 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c5ad4db-e797-4c07-885b-df111a70e0d0" containerName="collect-profiles" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.916108 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c5ad4db-e797-4c07-885b-df111a70e0d0" containerName="collect-profiles" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.916299 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c5ad4db-e797-4c07-885b-df111a70e0d0" containerName="collect-profiles" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.916935 4802 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.917154 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.917309 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36" gracePeriod=15 Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.917348 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094" gracePeriod=15 Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.917369 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80" gracePeriod=15 Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.917275 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875" gracePeriod=15 Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.917399 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488" gracePeriod=15 Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.917893 4802 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 17:15:32 crc kubenswrapper[4802]: E1125 17:15:32.918233 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.918254 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 17:15:32 crc kubenswrapper[4802]: E1125 17:15:32.918285 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.918294 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 17:15:32 crc kubenswrapper[4802]: E1125 17:15:32.918309 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.918316 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 17:15:32 crc kubenswrapper[4802]: E1125 17:15:32.918330 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.918339 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 17:15:32 crc kubenswrapper[4802]: E1125 17:15:32.918353 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.918361 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 17:15:32 crc kubenswrapper[4802]: E1125 17:15:32.918376 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.918382 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.918519 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.918536 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.918552 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.918562 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.918575 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.975115 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.975201 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.975346 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.975409 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.975448 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.975480 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.975581 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:32 crc kubenswrapper[4802]: I1125 17:15:32.975622 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:32 crc kubenswrapper[4802]: E1125 17:15:32.976756 4802 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.245:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.076208 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.076562 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.076603 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.076627 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.076682 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.076714 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.076766 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.076791 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.076862 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.076293 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.076913 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.076944 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.076974 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.077001 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.077031 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.077060 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.146686 4802 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.146743 4802 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Nov 25 17:15:33 crc kubenswrapper[4802]: E1125 17:15:33.147303 4802 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.245:6443: connect: connection refused" event=< Nov 25 17:15:33 crc kubenswrapper[4802]: &Event{ObjectMeta:{kube-apiserver-crc.187b4f5b6db2ad56 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Readiness probe error: Get "https://192.168.126.11:6443/readyz": dial tcp 192.168.126.11:6443: connect: connection refused Nov 25 17:15:33 crc kubenswrapper[4802]: body: Nov 25 17:15:33 crc kubenswrapper[4802]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 17:15:33.146725718 +0000 UTC m=+1716.291072904,LastTimestamp:2025-11-25 17:15:33.146725718 +0000 UTC m=+1716.291072904,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Nov 25 17:15:33 crc kubenswrapper[4802]: > Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.277354 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.827736 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"199b60b60e389b91cf44d85dd7d0a0f78d6af307f5090038b9d637fe177c2f15"} Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.827985 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"82431a2367feb87f7d24bded1ba920b23f662ef5055c6732931b8dd3fd7c8bcd"} Nov 25 17:15:33 crc kubenswrapper[4802]: E1125 17:15:33.828850 4802 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.245:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.830361 4802 generic.go:334] "Generic (PLEG): container finished" podID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" containerID="ce87fcfa6d3a6ade986c6bc2252d8dbc050b6b4ceb82c02d31b1612e9df4935c" exitCode=0 Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.830417 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7a7f65e4-cae3-49e6-8f88-df9a7e115260","Type":"ContainerDied","Data":"ce87fcfa6d3a6ade986c6bc2252d8dbc050b6b4ceb82c02d31b1612e9df4935c"} Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.831246 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.834423 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.835172 4802 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36" exitCode=0 Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.835191 4802 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094" exitCode=0 Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.835198 4802 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80" exitCode=0 Nov 25 17:15:33 crc kubenswrapper[4802]: I1125 17:15:33.835205 4802 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488" exitCode=2 Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.210503 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.211705 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.240418 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7a7f65e4-cae3-49e6-8f88-df9a7e115260-var-lock\") pod \"7a7f65e4-cae3-49e6-8f88-df9a7e115260\" (UID: \"7a7f65e4-cae3-49e6-8f88-df9a7e115260\") " Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.240516 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7a7f65e4-cae3-49e6-8f88-df9a7e115260-kubelet-dir\") pod \"7a7f65e4-cae3-49e6-8f88-df9a7e115260\" (UID: \"7a7f65e4-cae3-49e6-8f88-df9a7e115260\") " Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.240510 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a7f65e4-cae3-49e6-8f88-df9a7e115260-var-lock" (OuterVolumeSpecName: "var-lock") pod "7a7f65e4-cae3-49e6-8f88-df9a7e115260" (UID: "7a7f65e4-cae3-49e6-8f88-df9a7e115260"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.240612 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a7f65e4-cae3-49e6-8f88-df9a7e115260-kube-api-access\") pod \"7a7f65e4-cae3-49e6-8f88-df9a7e115260\" (UID: \"7a7f65e4-cae3-49e6-8f88-df9a7e115260\") " Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.240636 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a7f65e4-cae3-49e6-8f88-df9a7e115260-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7a7f65e4-cae3-49e6-8f88-df9a7e115260" (UID: "7a7f65e4-cae3-49e6-8f88-df9a7e115260"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.240956 4802 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7a7f65e4-cae3-49e6-8f88-df9a7e115260-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.240974 4802 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7a7f65e4-cae3-49e6-8f88-df9a7e115260-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.246249 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a7f65e4-cae3-49e6-8f88-df9a7e115260-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7a7f65e4-cae3-49e6-8f88-df9a7e115260" (UID: "7a7f65e4-cae3-49e6-8f88-df9a7e115260"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.343259 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a7f65e4-cae3-49e6-8f88-df9a7e115260-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.401422 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.402406 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.403058 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.403630 4802 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.444685 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.444903 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.444905 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.444941 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.444978 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.445082 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.445277 4802 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.445289 4802 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.445297 4802 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.533180 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.851574 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.852089 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7a7f65e4-cae3-49e6-8f88-df9a7e115260","Type":"ContainerDied","Data":"9fa31319a54e88d94c034ee5c4478d400fa958b65506fffc31ebb8dd3dbba150"} Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.852137 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9fa31319a54e88d94c034ee5c4478d400fa958b65506fffc31ebb8dd3dbba150" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.855873 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.856366 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.856902 4802 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875" exitCode=0 Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.856975 4802 scope.go:117] "RemoveContainer" containerID="bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.857318 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.857924 4802 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.858240 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.861999 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.862439 4802 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.873915 4802 scope.go:117] "RemoveContainer" containerID="dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.888246 4802 scope.go:117] "RemoveContainer" containerID="33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.908009 4802 scope.go:117] "RemoveContainer" containerID="2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.924949 4802 scope.go:117] "RemoveContainer" containerID="eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.943919 4802 scope.go:117] "RemoveContainer" containerID="3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.962916 4802 scope.go:117] "RemoveContainer" containerID="bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36" Nov 25 17:15:35 crc kubenswrapper[4802]: E1125 17:15:35.963344 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\": container with ID starting with bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36 not found: ID does not exist" containerID="bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.963384 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36"} err="failed to get container status \"bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\": rpc error: code = NotFound desc = could not find container \"bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36\": container with ID starting with bf039195d5de5e6d77bd99ec4898dd7d7deea9b6adb08921947d0b09e140ba36 not found: ID does not exist" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.963411 4802 scope.go:117] "RemoveContainer" containerID="dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094" Nov 25 17:15:35 crc kubenswrapper[4802]: E1125 17:15:35.963818 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\": container with ID starting with dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094 not found: ID does not exist" containerID="dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.963881 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094"} err="failed to get container status \"dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\": rpc error: code = NotFound desc = could not find container \"dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094\": container with ID starting with dbcb7a8c7c439be4e3ed52fe1e180c46e5f3e711d73e37c341cfe491ad635094 not found: ID does not exist" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.963922 4802 scope.go:117] "RemoveContainer" containerID="33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80" Nov 25 17:15:35 crc kubenswrapper[4802]: E1125 17:15:35.964453 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\": container with ID starting with 33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80 not found: ID does not exist" containerID="33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.964488 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80"} err="failed to get container status \"33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\": rpc error: code = NotFound desc = could not find container \"33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80\": container with ID starting with 33bc86a6025c16337884e96dd292971c4ca6939536434f013ce1267addd6fb80 not found: ID does not exist" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.964509 4802 scope.go:117] "RemoveContainer" containerID="2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488" Nov 25 17:15:35 crc kubenswrapper[4802]: E1125 17:15:35.964805 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\": container with ID starting with 2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488 not found: ID does not exist" containerID="2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.964852 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488"} err="failed to get container status \"2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\": rpc error: code = NotFound desc = could not find container \"2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488\": container with ID starting with 2419d8a50f28f245319817928693d2fb1cbbdb14a5669d33eacb41052e66b488 not found: ID does not exist" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.964867 4802 scope.go:117] "RemoveContainer" containerID="eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875" Nov 25 17:15:35 crc kubenswrapper[4802]: E1125 17:15:35.965247 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\": container with ID starting with eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875 not found: ID does not exist" containerID="eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.965273 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875"} err="failed to get container status \"eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\": rpc error: code = NotFound desc = could not find container \"eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875\": container with ID starting with eece4a1a6b9c82a2b79a3577e6ad58141fdffb524aca81233064acac6f03d875 not found: ID does not exist" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.965288 4802 scope.go:117] "RemoveContainer" containerID="3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375" Nov 25 17:15:35 crc kubenswrapper[4802]: E1125 17:15:35.965635 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\": container with ID starting with 3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375 not found: ID does not exist" containerID="3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375" Nov 25 17:15:35 crc kubenswrapper[4802]: I1125 17:15:35.965674 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375"} err="failed to get container status \"3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\": rpc error: code = NotFound desc = could not find container \"3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375\": container with ID starting with 3343e7de6e409478cf48b47dad1e3877d5c3bee73f8a43746ebf5790419d2375 not found: ID does not exist" Nov 25 17:15:36 crc kubenswrapper[4802]: E1125 17:15:36.810045 4802 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:36 crc kubenswrapper[4802]: E1125 17:15:36.810849 4802 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:36 crc kubenswrapper[4802]: E1125 17:15:36.811480 4802 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:36 crc kubenswrapper[4802]: E1125 17:15:36.811905 4802 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:36 crc kubenswrapper[4802]: E1125 17:15:36.812198 4802 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:36 crc kubenswrapper[4802]: I1125 17:15:36.812256 4802 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 25 17:15:36 crc kubenswrapper[4802]: E1125 17:15:36.812595 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" interval="200ms" Nov 25 17:15:37 crc kubenswrapper[4802]: E1125 17:15:37.014008 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" interval="400ms" Nov 25 17:15:37 crc kubenswrapper[4802]: E1125 17:15:37.415718 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" interval="800ms" Nov 25 17:15:37 crc kubenswrapper[4802]: I1125 17:15:37.514585 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:15:37 crc kubenswrapper[4802]: E1125 17:15:37.515026 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:15:37 crc kubenswrapper[4802]: I1125 17:15:37.516055 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:37 crc kubenswrapper[4802]: I1125 17:15:37.516610 4802 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:37 crc kubenswrapper[4802]: E1125 17:15:37.717404 4802 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.245:6443: connect: connection refused" event=< Nov 25 17:15:37 crc kubenswrapper[4802]: &Event{ObjectMeta:{kube-apiserver-crc.187b4f5b6db2ad56 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Readiness probe error: Get "https://192.168.126.11:6443/readyz": dial tcp 192.168.126.11:6443: connect: connection refused Nov 25 17:15:37 crc kubenswrapper[4802]: body: Nov 25 17:15:37 crc kubenswrapper[4802]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 17:15:33.146725718 +0000 UTC m=+1716.291072904,LastTimestamp:2025-11-25 17:15:33.146725718 +0000 UTC m=+1716.291072904,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Nov 25 17:15:37 crc kubenswrapper[4802]: > Nov 25 17:15:38 crc kubenswrapper[4802]: E1125 17:15:38.188675 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T17:15:38Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T17:15:38Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T17:15:38Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T17:15:38Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:38 crc kubenswrapper[4802]: E1125 17:15:38.189208 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:38 crc kubenswrapper[4802]: E1125 17:15:38.189530 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:38 crc kubenswrapper[4802]: E1125 17:15:38.189718 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:38 crc kubenswrapper[4802]: E1125 17:15:38.189897 4802 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:38 crc kubenswrapper[4802]: E1125 17:15:38.189910 4802 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 17:15:38 crc kubenswrapper[4802]: E1125 17:15:38.216531 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" interval="1.6s" Nov 25 17:15:38 crc kubenswrapper[4802]: E1125 17:15:38.581394 4802 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC glance-kuttl-tests/swift-swift-storage-0: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/persistentvolumeclaims/swift-swift-storage-0\": dial tcp 38.102.83.245:6443: connect: connection refused" pod="glance-kuttl-tests/swift-storage-0" volumeName="swift" Nov 25 17:15:39 crc kubenswrapper[4802]: E1125 17:15:39.817445 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" interval="3.2s" Nov 25 17:15:43 crc kubenswrapper[4802]: E1125 17:15:43.019075 4802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" interval="6.4s" Nov 25 17:15:44 crc kubenswrapper[4802]: E1125 17:15:44.568742 4802 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC glance-kuttl-tests/glance-glance-default-external-api-0: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/persistentvolumeclaims/glance-glance-default-external-api-0\": dial tcp 38.102.83.245:6443: connect: connection refused" pod="glance-kuttl-tests/glance-default-external-api-0" volumeName="glance" Nov 25 17:15:44 crc kubenswrapper[4802]: E1125 17:15:44.569272 4802 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC glance-kuttl-tests/glance-cache-glance-default-external-api-0: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/persistentvolumeclaims/glance-cache-glance-default-external-api-0\": dial tcp 38.102.83.245:6443: connect: connection refused" pod="glance-kuttl-tests/glance-default-external-api-0" volumeName="glance-cache" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.926189 4802 generic.go:334] "Generic (PLEG): container finished" podID="eeca972a-357e-4fa8-a8b5-1421b6601219" containerID="5fb925cd4ba59e71662af756f38cebc30afb1d1129d07ed4b4334801d2f89d4c" exitCode=1 Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.926295 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" event={"ID":"eeca972a-357e-4fa8-a8b5-1421b6601219","Type":"ContainerDied","Data":"5fb925cd4ba59e71662af756f38cebc30afb1d1129d07ed4b4334801d2f89d4c"} Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.927338 4802 scope.go:117] "RemoveContainer" containerID="5fb925cd4ba59e71662af756f38cebc30afb1d1129d07ed4b4334801d2f89d4c" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.927531 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.927974 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.928373 4802 generic.go:334] "Generic (PLEG): container finished" podID="a1724655-9ac1-43dc-8292-f39870b4a855" containerID="9d01284fca842bdc1379cf7add88c1dbce08a5591c1f4bb10ac4316ee241d9c5" exitCode=1 Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.928467 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" event={"ID":"a1724655-9ac1-43dc-8292-f39870b4a855","Type":"ContainerDied","Data":"9d01284fca842bdc1379cf7add88c1dbce08a5591c1f4bb10ac4316ee241d9c5"} Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.929809 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.929817 4802 scope.go:117] "RemoveContainer" containerID="9d01284fca842bdc1379cf7add88c1dbce08a5591c1f4bb10ac4316ee241d9c5" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.930141 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.930298 4802 generic.go:334] "Generic (PLEG): container finished" podID="c328497f-325f-4500-ba48-d1523cd7ee74" containerID="9861fdf30e2ca091950f978e9221e687d7cdd4a0c39d27cb97f49db9d42753fa" exitCode=1 Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.930350 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" event={"ID":"c328497f-325f-4500-ba48-d1523cd7ee74","Type":"ContainerDied","Data":"9861fdf30e2ca091950f978e9221e687d7cdd4a0c39d27cb97f49db9d42753fa"} Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.930498 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.930652 4802 scope.go:117] "RemoveContainer" containerID="9861fdf30e2ca091950f978e9221e687d7cdd4a0c39d27cb97f49db9d42753fa" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.930939 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.931349 4802 status_manager.go:851] "Failed to get status for pod" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-d58bc84f4-fjc7z\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.931913 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.932168 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.935470 4802 generic.go:334] "Generic (PLEG): container finished" podID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" containerID="7e4484cf13bf5fb149873ab54dc09699f84f1df1654eb8e4997afa2fb919a339" exitCode=1 Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.935534 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" event={"ID":"373fd8b2-c469-46b8-b20e-8ecd875c4a39","Type":"ContainerDied","Data":"7e4484cf13bf5fb149873ab54dc09699f84f1df1654eb8e4997afa2fb919a339"} Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.935903 4802 scope.go:117] "RemoveContainer" containerID="7e4484cf13bf5fb149873ab54dc09699f84f1df1654eb8e4997afa2fb919a339" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.937688 4802 status_manager.go:851] "Failed to get status for pod" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-d58bc84f4-fjc7z\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.938070 4802 generic.go:334] "Generic (PLEG): container finished" podID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" containerID="342193f4dc464f0cf86d641c2c4e239dbe6d8a0ae54e1e1ef2fb90de1debc0d4" exitCode=1 Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.938151 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" event={"ID":"3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb","Type":"ContainerDied","Data":"342193f4dc464f0cf86d641c2c4e239dbe6d8a0ae54e1e1ef2fb90de1debc0d4"} Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.939755 4802 scope.go:117] "RemoveContainer" containerID="342193f4dc464f0cf86d641c2c4e239dbe6d8a0ae54e1e1ef2fb90de1debc0d4" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.940210 4802 generic.go:334] "Generic (PLEG): container finished" podID="a5bdabd0-7540-4d45-8675-9bd88c415957" containerID="c047bba0964146b17129c401fb29cd7d6ccbedf4efe035eab267e9821dfc0d58" exitCode=1 Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.940283 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" event={"ID":"a5bdabd0-7540-4d45-8675-9bd88c415957","Type":"ContainerDied","Data":"c047bba0964146b17129c401fb29cd7d6ccbedf4efe035eab267e9821dfc0d58"} Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.940724 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.940875 4802 scope.go:117] "RemoveContainer" containerID="c047bba0964146b17129c401fb29cd7d6ccbedf4efe035eab267e9821dfc0d58" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.941438 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.941953 4802 status_manager.go:851] "Failed to get status for pod" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-987nc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.942471 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.942924 4802 status_manager.go:851] "Failed to get status for pod" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-d58bc84f4-fjc7z\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.943009 4802 generic.go:334] "Generic (PLEG): container finished" podID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" containerID="5fdb255c3eda15bc7a4f2954f1a8b55564ad5bf1940a9e1963595f5ec97e4742" exitCode=1 Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.943065 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" event={"ID":"eda71d4e-8d95-4944-a0b8-a8ac84c01ca4","Type":"ContainerDied","Data":"5fdb255c3eda15bc7a4f2954f1a8b55564ad5bf1940a9e1963595f5ec97e4742"} Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.943235 4802 status_manager.go:851] "Failed to get status for pod" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-65c6fdb94d-5cfbr\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.943441 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.943656 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.943932 4802 status_manager.go:851] "Failed to get status for pod" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-987nc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.944567 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.944792 4802 status_manager.go:851] "Failed to get status for pod" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-5787d846c-g6b4c\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.945042 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.945269 4802 status_manager.go:851] "Failed to get status for pod" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-987nc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.945450 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.945645 4802 status_manager.go:851] "Failed to get status for pod" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-5787d846c-g6b4c\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.945869 4802 status_manager.go:851] "Failed to get status for pod" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-d58bc84f4-fjc7z\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.946243 4802 status_manager.go:851] "Failed to get status for pod" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6fccf5cdb-rrwcf\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.946296 4802 generic.go:334] "Generic (PLEG): container finished" podID="5004bc39-3a06-4842-9565-8501f35c8c4b" containerID="f36e88039dd0e0ed5ed95ac4f1f0d31d7927e4940038a72419458c2d43e546d4" exitCode=1 Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.946320 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" event={"ID":"5004bc39-3a06-4842-9565-8501f35c8c4b","Type":"ContainerDied","Data":"f36e88039dd0e0ed5ed95ac4f1f0d31d7927e4940038a72419458c2d43e546d4"} Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.946449 4802 scope.go:117] "RemoveContainer" containerID="5fdb255c3eda15bc7a4f2954f1a8b55564ad5bf1940a9e1963595f5ec97e4742" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.946580 4802 status_manager.go:851] "Failed to get status for pod" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-65c6fdb94d-5cfbr\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.946609 4802 scope.go:117] "RemoveContainer" containerID="f36e88039dd0e0ed5ed95ac4f1f0d31d7927e4940038a72419458c2d43e546d4" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.946785 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.947591 4802 status_manager.go:851] "Failed to get status for pod" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-5787d846c-g6b4c\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.950737 4802 status_manager.go:851] "Failed to get status for pod" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-5959d99654-4j2qc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.951684 4802 status_manager.go:851] "Failed to get status for pod" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-d58bc84f4-fjc7z\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.952035 4802 status_manager.go:851] "Failed to get status for pod" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6fccf5cdb-rrwcf\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.952529 4802 status_manager.go:851] "Failed to get status for pod" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-65c6fdb94d-5cfbr\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.952975 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.953450 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.953919 4802 status_manager.go:851] "Failed to get status for pod" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-987nc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:44 crc kubenswrapper[4802]: I1125 17:15:44.954189 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.290911 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.846398 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.846845 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.958645 4802 generic.go:334] "Generic (PLEG): container finished" podID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" containerID="f59fe8a229e461bdaeef0cc7495dc1200a263f9210f6e3aaf0c2a051ac50c485" exitCode=1 Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.958785 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" event={"ID":"373fd8b2-c469-46b8-b20e-8ecd875c4a39","Type":"ContainerDied","Data":"f59fe8a229e461bdaeef0cc7495dc1200a263f9210f6e3aaf0c2a051ac50c485"} Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.958877 4802 scope.go:117] "RemoveContainer" containerID="7e4484cf13bf5fb149873ab54dc09699f84f1df1654eb8e4997afa2fb919a339" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.959592 4802 scope.go:117] "RemoveContainer" containerID="f59fe8a229e461bdaeef0cc7495dc1200a263f9210f6e3aaf0c2a051ac50c485" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.959961 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: E1125 17:15:45.960074 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=operator pod=rabbitmq-cluster-operator-779fc9694b-987nc_openstack-operators(373fd8b2-c469-46b8-b20e-8ecd875c4a39)\"" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.960672 4802 status_manager.go:851] "Failed to get status for pod" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-5787d846c-g6b4c\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.961178 4802 status_manager.go:851] "Failed to get status for pod" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-5959d99654-4j2qc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.961367 4802 generic.go:334] "Generic (PLEG): container finished" podID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" containerID="93ea15cbbe4d99268f06277f2d93ce56c43989561de09ed1af2238675c896eac" exitCode=1 Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.961449 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" event={"ID":"3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb","Type":"ContainerDied","Data":"93ea15cbbe4d99268f06277f2d93ce56c43989561de09ed1af2238675c896eac"} Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.962007 4802 status_manager.go:851] "Failed to get status for pod" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-d58bc84f4-fjc7z\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.962389 4802 scope.go:117] "RemoveContainer" containerID="93ea15cbbe4d99268f06277f2d93ce56c43989561de09ed1af2238675c896eac" Nov 25 17:15:45 crc kubenswrapper[4802]: E1125 17:15:45.962807 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=swift-operator-controller-manager-5787d846c-g6b4c_openstack-operators(3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb)\"" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.962964 4802 status_manager.go:851] "Failed to get status for pod" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6fccf5cdb-rrwcf\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.963560 4802 status_manager.go:851] "Failed to get status for pod" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-65c6fdb94d-5cfbr\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.964038 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.964521 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.964735 4802 generic.go:334] "Generic (PLEG): container finished" podID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" containerID="9ce5ad58c88b39008d469294bae4d4e6c1de715142944b1387e62b498caf2b4f" exitCode=1 Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.964835 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" event={"ID":"eda71d4e-8d95-4944-a0b8-a8ac84c01ca4","Type":"ContainerDied","Data":"9ce5ad58c88b39008d469294bae4d4e6c1de715142944b1387e62b498caf2b4f"} Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.964976 4802 status_manager.go:851] "Failed to get status for pod" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-987nc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.965702 4802 status_manager.go:851] "Failed to get status for pod" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-d58bc84f4-fjc7z\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.965931 4802 scope.go:117] "RemoveContainer" containerID="9ce5ad58c88b39008d469294bae4d4e6c1de715142944b1387e62b498caf2b4f" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.966353 4802 status_manager.go:851] "Failed to get status for pod" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6fccf5cdb-rrwcf\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: E1125 17:15:45.966407 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=infra-operator-controller-manager-6fccf5cdb-rrwcf_openstack-operators(eda71d4e-8d95-4944-a0b8-a8ac84c01ca4)\"" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.967040 4802 generic.go:334] "Generic (PLEG): container finished" podID="c328497f-325f-4500-ba48-d1523cd7ee74" containerID="e903f73446fc13ae8e0c35edd104f75f1b890874642288c1dc68bfe6345bc27e" exitCode=1 Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.967156 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" event={"ID":"c328497f-325f-4500-ba48-d1523cd7ee74","Type":"ContainerDied","Data":"e903f73446fc13ae8e0c35edd104f75f1b890874642288c1dc68bfe6345bc27e"} Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.967149 4802 status_manager.go:851] "Failed to get status for pod" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-65c6fdb94d-5cfbr\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.967638 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.967841 4802 scope.go:117] "RemoveContainer" containerID="e903f73446fc13ae8e0c35edd104f75f1b890874642288c1dc68bfe6345bc27e" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.968083 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: E1125 17:15:45.968321 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=horizon-operator-controller-manager-d58bc84f4-fjc7z_openstack-operators(c328497f-325f-4500-ba48-d1523cd7ee74)\"" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.968568 4802 status_manager.go:851] "Failed to get status for pod" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-987nc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.969114 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.969713 4802 status_manager.go:851] "Failed to get status for pod" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-5787d846c-g6b4c\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.970244 4802 status_manager.go:851] "Failed to get status for pod" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-5959d99654-4j2qc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.970905 4802 status_manager.go:851] "Failed to get status for pod" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-d58bc84f4-fjc7z\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.971451 4802 status_manager.go:851] "Failed to get status for pod" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6fccf5cdb-rrwcf\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.971955 4802 status_manager.go:851] "Failed to get status for pod" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-65c6fdb94d-5cfbr\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.972508 4802 generic.go:334] "Generic (PLEG): container finished" podID="5004bc39-3a06-4842-9565-8501f35c8c4b" containerID="0fe44caae2012d33b6d3e7f490231a1b33fdf67b9937e39c9568563ae4d863e4" exitCode=1 Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.972595 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" event={"ID":"5004bc39-3a06-4842-9565-8501f35c8c4b","Type":"ContainerDied","Data":"0fe44caae2012d33b6d3e7f490231a1b33fdf67b9937e39c9568563ae4d863e4"} Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.972674 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.973026 4802 scope.go:117] "RemoveContainer" containerID="0fe44caae2012d33b6d3e7f490231a1b33fdf67b9937e39c9568563ae4d863e4" Nov 25 17:15:45 crc kubenswrapper[4802]: E1125 17:15:45.973318 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=mariadb-operator-controller-manager-5959d99654-4j2qc_openstack-operators(5004bc39-3a06-4842-9565-8501f35c8c4b)\"" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.973425 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.973827 4802 status_manager.go:851] "Failed to get status for pod" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-987nc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.974202 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.974575 4802 status_manager.go:851] "Failed to get status for pod" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-5787d846c-g6b4c\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.974896 4802 status_manager.go:851] "Failed to get status for pod" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-5959d99654-4j2qc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.975360 4802 status_manager.go:851] "Failed to get status for pod" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-5787d846c-g6b4c\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.975442 4802 generic.go:334] "Generic (PLEG): container finished" podID="eeca972a-357e-4fa8-a8b5-1421b6601219" containerID="85bb3095155adb4276be27c6191c3475eb3bba0054184eac5e1cb2669bd15a66" exitCode=1 Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.975527 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" event={"ID":"eeca972a-357e-4fa8-a8b5-1421b6601219","Type":"ContainerDied","Data":"85bb3095155adb4276be27c6191c3475eb3bba0054184eac5e1cb2669bd15a66"} Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.975696 4802 status_manager.go:851] "Failed to get status for pod" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-5959d99654-4j2qc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.975929 4802 status_manager.go:851] "Failed to get status for pod" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-d58bc84f4-fjc7z\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.976141 4802 status_manager.go:851] "Failed to get status for pod" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6fccf5cdb-rrwcf\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.976208 4802 scope.go:117] "RemoveContainer" containerID="85bb3095155adb4276be27c6191c3475eb3bba0054184eac5e1cb2669bd15a66" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.976412 4802 status_manager.go:851] "Failed to get status for pod" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-65c6fdb94d-5cfbr\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: E1125 17:15:45.976516 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=glance-operator-controller-manager-6bfbf5c856-xw6fj_openstack-operators(eeca972a-357e-4fa8-a8b5-1421b6601219)\"" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.976729 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.977030 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.977360 4802 status_manager.go:851] "Failed to get status for pod" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-987nc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.977700 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.978205 4802 status_manager.go:851] "Failed to get status for pod" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-d58bc84f4-fjc7z\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.978390 4802 generic.go:334] "Generic (PLEG): container finished" podID="a1724655-9ac1-43dc-8292-f39870b4a855" containerID="1dddde497d38cccc79697be659e98a97c142049481be6f0e11bc8312d99da3af" exitCode=1 Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.978419 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" event={"ID":"a1724655-9ac1-43dc-8292-f39870b4a855","Type":"ContainerDied","Data":"1dddde497d38cccc79697be659e98a97c142049481be6f0e11bc8312d99da3af"} Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.978597 4802 status_manager.go:851] "Failed to get status for pod" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6fccf5cdb-rrwcf\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.978959 4802 status_manager.go:851] "Failed to get status for pod" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-65c6fdb94d-5cfbr\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.979240 4802 scope.go:117] "RemoveContainer" containerID="1dddde497d38cccc79697be659e98a97c142049481be6f0e11bc8312d99da3af" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.979263 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: E1125 17:15:45.979608 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=metallb-operator-controller-manager-548f76d7c-s5g2f_metallb-system(a1724655-9ac1-43dc-8292-f39870b4a855)\"" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.979636 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.979976 4802 status_manager.go:851] "Failed to get status for pod" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-987nc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.980232 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.980520 4802 status_manager.go:851] "Failed to get status for pod" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-5787d846c-g6b4c\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.980727 4802 status_manager.go:851] "Failed to get status for pod" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-5959d99654-4j2qc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.981077 4802 status_manager.go:851] "Failed to get status for pod" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-5787d846c-g6b4c\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.981424 4802 generic.go:334] "Generic (PLEG): container finished" podID="a5bdabd0-7540-4d45-8675-9bd88c415957" containerID="b937b9f3a5d081aa5c27c6bff3a0ac96523739001944f058e149959f1d8068dc" exitCode=1 Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.981455 4802 status_manager.go:851] "Failed to get status for pod" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-5959d99654-4j2qc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.981485 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" event={"ID":"a5bdabd0-7540-4d45-8675-9bd88c415957","Type":"ContainerDied","Data":"b937b9f3a5d081aa5c27c6bff3a0ac96523739001944f058e149959f1d8068dc"} Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.981978 4802 scope.go:117] "RemoveContainer" containerID="b937b9f3a5d081aa5c27c6bff3a0ac96523739001944f058e149959f1d8068dc" Nov 25 17:15:45 crc kubenswrapper[4802]: E1125 17:15:45.982212 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-65c6fdb94d-5cfbr_openstack-operators(a5bdabd0-7540-4d45-8675-9bd88c415957)\"" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.982311 4802 status_manager.go:851] "Failed to get status for pod" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-d58bc84f4-fjc7z\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.982558 4802 status_manager.go:851] "Failed to get status for pod" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6fccf5cdb-rrwcf\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.982871 4802 status_manager.go:851] "Failed to get status for pod" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-65c6fdb94d-5cfbr\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.983106 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.983438 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.983639 4802 status_manager.go:851] "Failed to get status for pod" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-987nc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.983799 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.983995 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.984270 4802 status_manager.go:851] "Failed to get status for pod" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-5787d846c-g6b4c\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.984651 4802 status_manager.go:851] "Failed to get status for pod" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-5959d99654-4j2qc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.984978 4802 status_manager.go:851] "Failed to get status for pod" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-d58bc84f4-fjc7z\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.985288 4802 status_manager.go:851] "Failed to get status for pod" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6fccf5cdb-rrwcf\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.985809 4802 status_manager.go:851] "Failed to get status for pod" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-65c6fdb94d-5cfbr\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.986363 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.986470 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.986526 4802 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89" exitCode=1 Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.986563 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89"} Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.986664 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.986930 4802 status_manager.go:851] "Failed to get status for pod" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-987nc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.987275 4802 scope.go:117] "RemoveContainer" containerID="bba48bb2fc069baea1ba95f418f9e4f8690a1a81e1cde26559091d96e53e7e89" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.987498 4802 status_manager.go:851] "Failed to get status for pod" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-d58bc84f4-fjc7z\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.987841 4802 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.988145 4802 status_manager.go:851] "Failed to get status for pod" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6fccf5cdb-rrwcf\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.988557 4802 status_manager.go:851] "Failed to get status for pod" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-65c6fdb94d-5cfbr\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.989007 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.989485 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.989776 4802 status_manager.go:851] "Failed to get status for pod" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-987nc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.990155 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.990478 4802 status_manager.go:851] "Failed to get status for pod" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-5787d846c-g6b4c\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:45 crc kubenswrapper[4802]: I1125 17:15:45.991092 4802 status_manager.go:851] "Failed to get status for pod" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-5959d99654-4j2qc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.006913 4802 scope.go:117] "RemoveContainer" containerID="342193f4dc464f0cf86d641c2c4e239dbe6d8a0ae54e1e1ef2fb90de1debc0d4" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.087959 4802 scope.go:117] "RemoveContainer" containerID="5fdb255c3eda15bc7a4f2954f1a8b55564ad5bf1940a9e1963595f5ec97e4742" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.127816 4802 scope.go:117] "RemoveContainer" containerID="9861fdf30e2ca091950f978e9221e687d7cdd4a0c39d27cb97f49db9d42753fa" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.164772 4802 scope.go:117] "RemoveContainer" containerID="f36e88039dd0e0ed5ed95ac4f1f0d31d7927e4940038a72419458c2d43e546d4" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.215864 4802 scope.go:117] "RemoveContainer" containerID="5fb925cd4ba59e71662af756f38cebc30afb1d1129d07ed4b4334801d2f89d4c" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.239233 4802 scope.go:117] "RemoveContainer" containerID="9d01284fca842bdc1379cf7add88c1dbce08a5591c1f4bb10ac4316ee241d9c5" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.258292 4802 scope.go:117] "RemoveContainer" containerID="c047bba0964146b17129c401fb29cd7d6ccbedf4efe035eab267e9821dfc0d58" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.312282 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.504461 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.505371 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.505699 4802 status_manager.go:851] "Failed to get status for pod" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-987nc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.506067 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.506286 4802 status_manager.go:851] "Failed to get status for pod" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-5787d846c-g6b4c\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.506473 4802 status_manager.go:851] "Failed to get status for pod" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-5959d99654-4j2qc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.506742 4802 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.507013 4802 status_manager.go:851] "Failed to get status for pod" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-d58bc84f4-fjc7z\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.507422 4802 status_manager.go:851] "Failed to get status for pod" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6fccf5cdb-rrwcf\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.510676 4802 status_manager.go:851] "Failed to get status for pod" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-65c6fdb94d-5cfbr\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.511016 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.520172 4802 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9435c266-8324-4e76-bdc6-a0a63ab94230" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.520222 4802 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9435c266-8324-4e76-bdc6-a0a63ab94230" Nov 25 17:15:46 crc kubenswrapper[4802]: E1125 17:15:46.520662 4802 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:46 crc kubenswrapper[4802]: I1125 17:15:46.521139 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:46 crc kubenswrapper[4802]: W1125 17:15:46.544427 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-47491e813333ec7eb221b8cf54181935b0f2409e1fe6c3de96bae4763c2300a7 WatchSource:0}: Error finding container 47491e813333ec7eb221b8cf54181935b0f2409e1fe6c3de96bae4763c2300a7: Status 404 returned error can't find the container with id 47491e813333ec7eb221b8cf54181935b0f2409e1fe6c3de96bae4763c2300a7 Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.003651 4802 status_manager.go:851] "Failed to get status for pod" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-d58bc84f4-fjc7z\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.004066 4802 scope.go:117] "RemoveContainer" containerID="1dddde497d38cccc79697be659e98a97c142049481be6f0e11bc8312d99da3af" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.004372 4802 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: E1125 17:15:47.004535 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=metallb-operator-controller-manager-548f76d7c-s5g2f_metallb-system(a1724655-9ac1-43dc-8292-f39870b4a855)\"" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.004615 4802 status_manager.go:851] "Failed to get status for pod" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6fccf5cdb-rrwcf\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.005201 4802 status_manager.go:851] "Failed to get status for pod" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-65c6fdb94d-5cfbr\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.006041 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.006815 4802 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="ce7603d9e616e120551c66b66fcad4f628c54abff0eff550b2bbba6eebc1e61c" exitCode=0 Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.006892 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"ce7603d9e616e120551c66b66fcad4f628c54abff0eff550b2bbba6eebc1e61c"} Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.006929 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"47491e813333ec7eb221b8cf54181935b0f2409e1fe6c3de96bae4763c2300a7"} Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.007118 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.007226 4802 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9435c266-8324-4e76-bdc6-a0a63ab94230" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.007372 4802 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9435c266-8324-4e76-bdc6-a0a63ab94230" Nov 25 17:15:47 crc kubenswrapper[4802]: E1125 17:15:47.008289 4802 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.008571 4802 scope.go:117] "RemoveContainer" containerID="93ea15cbbe4d99268f06277f2d93ce56c43989561de09ed1af2238675c896eac" Nov 25 17:15:47 crc kubenswrapper[4802]: E1125 17:15:47.008777 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=swift-operator-controller-manager-5787d846c-g6b4c_openstack-operators(3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb)\"" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.009171 4802 status_manager.go:851] "Failed to get status for pod" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-987nc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.010751 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.011296 4802 status_manager.go:851] "Failed to get status for pod" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-5787d846c-g6b4c\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.011641 4802 status_manager.go:851] "Failed to get status for pod" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-5959d99654-4j2qc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.012280 4802 status_manager.go:851] "Failed to get status for pod" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-d58bc84f4-fjc7z\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.012546 4802 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.012770 4802 status_manager.go:851] "Failed to get status for pod" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6fccf5cdb-rrwcf\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.013059 4802 status_manager.go:851] "Failed to get status for pod" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-65c6fdb94d-5cfbr\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.013532 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.013940 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.014245 4802 status_manager.go:851] "Failed to get status for pod" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-987nc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.014588 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.014925 4802 status_manager.go:851] "Failed to get status for pod" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-5787d846c-g6b4c\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.015216 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.015317 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f6893dc5d5810030ac3f56de0bf24d02c03025701398aa5b1df9f37c67dfb1e6"} Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.015540 4802 status_manager.go:851] "Failed to get status for pod" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-5959d99654-4j2qc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.016262 4802 status_manager.go:851] "Failed to get status for pod" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-d58bc84f4-fjc7z\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.017666 4802 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.018009 4802 status_manager.go:851] "Failed to get status for pod" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6fccf5cdb-rrwcf\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.018446 4802 status_manager.go:851] "Failed to get status for pod" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-65c6fdb94d-5cfbr\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.018900 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.019352 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.019697 4802 status_manager.go:851] "Failed to get status for pod" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-987nc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.019932 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.020282 4802 status_manager.go:851] "Failed to get status for pod" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-5787d846c-g6b4c\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.020352 4802 scope.go:117] "RemoveContainer" containerID="0fe44caae2012d33b6d3e7f490231a1b33fdf67b9937e39c9568563ae4d863e4" Nov 25 17:15:47 crc kubenswrapper[4802]: E1125 17:15:47.020568 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=mariadb-operator-controller-manager-5959d99654-4j2qc_openstack-operators(5004bc39-3a06-4842-9565-8501f35c8c4b)\"" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.020680 4802 status_manager.go:851] "Failed to get status for pod" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-5959d99654-4j2qc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.020997 4802 status_manager.go:851] "Failed to get status for pod" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-5787d846c-g6b4c\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.021230 4802 status_manager.go:851] "Failed to get status for pod" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-5959d99654-4j2qc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.021462 4802 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.021776 4802 status_manager.go:851] "Failed to get status for pod" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-d58bc84f4-fjc7z\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.022096 4802 status_manager.go:851] "Failed to get status for pod" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-6fccf5cdb-rrwcf\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.022415 4802 status_manager.go:851] "Failed to get status for pod" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-65c6fdb94d-5cfbr\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.022714 4802 status_manager.go:851] "Failed to get status for pod" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-6bfbf5c856-xw6fj\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.023011 4802 status_manager.go:851] "Failed to get status for pod" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.023274 4802 status_manager.go:851] "Failed to get status for pod" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-987nc\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:47 crc kubenswrapper[4802]: I1125 17:15:47.023567 4802 status_manager.go:851] "Failed to get status for pod" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-548f76d7c-s5g2f\": dial tcp 38.102.83.245:6443: connect: connection refused" Nov 25 17:15:48 crc kubenswrapper[4802]: I1125 17:15:48.034701 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1c7be6583c0a6df60478fa1b9520e5352a4ce11503ee3fcabf56babc271f02c0"} Nov 25 17:15:48 crc kubenswrapper[4802]: I1125 17:15:48.035087 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"40d05d2b0aef4dca91c42794cb8d9325d12339b6350370d1d28e950fcd4393a7"} Nov 25 17:15:48 crc kubenswrapper[4802]: I1125 17:15:48.035109 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"83633034f72348ed5adc58c185c69d28d167639cb0d08bea9532aa8deecad395"} Nov 25 17:15:48 crc kubenswrapper[4802]: I1125 17:15:48.035180 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ef8835196774670e419eca2daeb972ca15476f30db6f7235dded159abb3c3579"} Nov 25 17:15:48 crc kubenswrapper[4802]: I1125 17:15:48.407965 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:15:48 crc kubenswrapper[4802]: I1125 17:15:48.408304 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:15:48 crc kubenswrapper[4802]: I1125 17:15:48.408570 4802 scope.go:117] "RemoveContainer" containerID="9ce5ad58c88b39008d469294bae4d4e6c1de715142944b1387e62b498caf2b4f" Nov 25 17:15:48 crc kubenswrapper[4802]: E1125 17:15:48.408819 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=infra-operator-controller-manager-6fccf5cdb-rrwcf_openstack-operators(eda71d4e-8d95-4944-a0b8-a8ac84c01ca4)\"" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" Nov 25 17:15:49 crc kubenswrapper[4802]: I1125 17:15:49.045381 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2d6af71d7bbdc66f6fd8773f7ab82c44bb0c2981f9a58f8d68c2d3c373d96d9a"} Nov 25 17:15:49 crc kubenswrapper[4802]: I1125 17:15:49.045578 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:49 crc kubenswrapper[4802]: I1125 17:15:49.045746 4802 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9435c266-8324-4e76-bdc6-a0a63ab94230" Nov 25 17:15:49 crc kubenswrapper[4802]: I1125 17:15:49.045769 4802 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9435c266-8324-4e76-bdc6-a0a63ab94230" Nov 25 17:15:49 crc kubenswrapper[4802]: I1125 17:15:49.045829 4802 scope.go:117] "RemoveContainer" containerID="9ce5ad58c88b39008d469294bae4d4e6c1de715142944b1387e62b498caf2b4f" Nov 25 17:15:49 crc kubenswrapper[4802]: E1125 17:15:49.046050 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=infra-operator-controller-manager-6fccf5cdb-rrwcf_openstack-operators(eda71d4e-8d95-4944-a0b8-a8ac84c01ca4)\"" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" Nov 25 17:15:49 crc kubenswrapper[4802]: I1125 17:15:49.105905 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:15:49 crc kubenswrapper[4802]: I1125 17:15:49.105955 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:15:49 crc kubenswrapper[4802]: I1125 17:15:49.106604 4802 scope.go:117] "RemoveContainer" containerID="b937b9f3a5d081aa5c27c6bff3a0ac96523739001944f058e149959f1d8068dc" Nov 25 17:15:49 crc kubenswrapper[4802]: E1125 17:15:49.106861 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-65c6fdb94d-5cfbr_openstack-operators(a5bdabd0-7540-4d45-8675-9bd88c415957)\"" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" Nov 25 17:15:49 crc kubenswrapper[4802]: I1125 17:15:49.199475 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:15:49 crc kubenswrapper[4802]: I1125 17:15:49.200288 4802 scope.go:117] "RemoveContainer" containerID="85bb3095155adb4276be27c6191c3475eb3bba0054184eac5e1cb2669bd15a66" Nov 25 17:15:49 crc kubenswrapper[4802]: E1125 17:15:49.200542 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=glance-operator-controller-manager-6bfbf5c856-xw6fj_openstack-operators(eeca972a-357e-4fa8-a8b5-1421b6601219)\"" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" Nov 25 17:15:51 crc kubenswrapper[4802]: I1125 17:15:51.522037 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:51 crc kubenswrapper[4802]: I1125 17:15:51.522293 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:51 crc kubenswrapper[4802]: I1125 17:15:51.527737 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:51 crc kubenswrapper[4802]: I1125 17:15:51.705703 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 17:15:52 crc kubenswrapper[4802]: I1125 17:15:52.354877 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:15:52 crc kubenswrapper[4802]: I1125 17:15:52.355841 4802 scope.go:117] "RemoveContainer" containerID="e903f73446fc13ae8e0c35edd104f75f1b890874642288c1dc68bfe6345bc27e" Nov 25 17:15:52 crc kubenswrapper[4802]: E1125 17:15:52.356161 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=horizon-operator-controller-manager-d58bc84f4-fjc7z_openstack-operators(c328497f-325f-4500-ba48-d1523cd7ee74)\"" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" Nov 25 17:15:52 crc kubenswrapper[4802]: I1125 17:15:52.505371 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:15:52 crc kubenswrapper[4802]: E1125 17:15:52.505598 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:15:53 crc kubenswrapper[4802]: I1125 17:15:53.148314 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 17:15:53 crc kubenswrapper[4802]: I1125 17:15:53.152445 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 17:15:54 crc kubenswrapper[4802]: I1125 17:15:54.053017 4802 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:54 crc kubenswrapper[4802]: I1125 17:15:54.084294 4802 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9435c266-8324-4e76-bdc6-a0a63ab94230" Nov 25 17:15:54 crc kubenswrapper[4802]: I1125 17:15:54.084324 4802 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9435c266-8324-4e76-bdc6-a0a63ab94230" Nov 25 17:15:54 crc kubenswrapper[4802]: I1125 17:15:54.088000 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:15:54 crc kubenswrapper[4802]: I1125 17:15:54.090581 4802 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="741e10ea-3677-4637-bb6b-fb42e05a58c8" Nov 25 17:15:55 crc kubenswrapper[4802]: I1125 17:15:55.092809 4802 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9435c266-8324-4e76-bdc6-a0a63ab94230" Nov 25 17:15:55 crc kubenswrapper[4802]: I1125 17:15:55.093295 4802 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9435c266-8324-4e76-bdc6-a0a63ab94230" Nov 25 17:15:55 crc kubenswrapper[4802]: I1125 17:15:55.290290 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:15:55 crc kubenswrapper[4802]: I1125 17:15:55.290536 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:15:55 crc kubenswrapper[4802]: I1125 17:15:55.291114 4802 scope.go:117] "RemoveContainer" containerID="93ea15cbbe4d99268f06277f2d93ce56c43989561de09ed1af2238675c896eac" Nov 25 17:15:55 crc kubenswrapper[4802]: E1125 17:15:55.291325 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=swift-operator-controller-manager-5787d846c-g6b4c_openstack-operators(3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb)\"" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" Nov 25 17:15:55 crc kubenswrapper[4802]: I1125 17:15:55.846739 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:15:55 crc kubenswrapper[4802]: I1125 17:15:55.847435 4802 scope.go:117] "RemoveContainer" containerID="0fe44caae2012d33b6d3e7f490231a1b33fdf67b9937e39c9568563ae4d863e4" Nov 25 17:15:56 crc kubenswrapper[4802]: I1125 17:15:56.104167 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" event={"ID":"5004bc39-3a06-4842-9565-8501f35c8c4b","Type":"ContainerStarted","Data":"16316a73afcdc8c5e68f8dfba330f336394584b43d0a454011d7380e169e24f1"} Nov 25 17:15:56 crc kubenswrapper[4802]: I1125 17:15:56.105681 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:15:57 crc kubenswrapper[4802]: I1125 17:15:57.114891 4802 generic.go:334] "Generic (PLEG): container finished" podID="5004bc39-3a06-4842-9565-8501f35c8c4b" containerID="16316a73afcdc8c5e68f8dfba330f336394584b43d0a454011d7380e169e24f1" exitCode=1 Nov 25 17:15:57 crc kubenswrapper[4802]: I1125 17:15:57.114943 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" event={"ID":"5004bc39-3a06-4842-9565-8501f35c8c4b","Type":"ContainerDied","Data":"16316a73afcdc8c5e68f8dfba330f336394584b43d0a454011d7380e169e24f1"} Nov 25 17:15:57 crc kubenswrapper[4802]: I1125 17:15:57.114984 4802 scope.go:117] "RemoveContainer" containerID="0fe44caae2012d33b6d3e7f490231a1b33fdf67b9937e39c9568563ae4d863e4" Nov 25 17:15:57 crc kubenswrapper[4802]: I1125 17:15:57.115373 4802 scope.go:117] "RemoveContainer" containerID="16316a73afcdc8c5e68f8dfba330f336394584b43d0a454011d7380e169e24f1" Nov 25 17:15:57 crc kubenswrapper[4802]: E1125 17:15:57.115589 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=mariadb-operator-controller-manager-5959d99654-4j2qc_openstack-operators(5004bc39-3a06-4842-9565-8501f35c8c4b)\"" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" Nov 25 17:15:57 crc kubenswrapper[4802]: I1125 17:15:57.563995 4802 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="741e10ea-3677-4637-bb6b-fb42e05a58c8" Nov 25 17:15:58 crc kubenswrapper[4802]: I1125 17:15:58.124072 4802 scope.go:117] "RemoveContainer" containerID="16316a73afcdc8c5e68f8dfba330f336394584b43d0a454011d7380e169e24f1" Nov 25 17:15:58 crc kubenswrapper[4802]: E1125 17:15:58.124654 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=mariadb-operator-controller-manager-5959d99654-4j2qc_openstack-operators(5004bc39-3a06-4842-9565-8501f35c8c4b)\"" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" Nov 25 17:15:59 crc kubenswrapper[4802]: I1125 17:15:59.199795 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:15:59 crc kubenswrapper[4802]: I1125 17:15:59.201494 4802 scope.go:117] "RemoveContainer" containerID="85bb3095155adb4276be27c6191c3475eb3bba0054184eac5e1cb2669bd15a66" Nov 25 17:15:59 crc kubenswrapper[4802]: I1125 17:15:59.505450 4802 scope.go:117] "RemoveContainer" containerID="f59fe8a229e461bdaeef0cc7495dc1200a263f9210f6e3aaf0c2a051ac50c485" Nov 25 17:16:00 crc kubenswrapper[4802]: I1125 17:16:00.147381 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" event={"ID":"373fd8b2-c469-46b8-b20e-8ecd875c4a39","Type":"ContainerStarted","Data":"a2251ba467ce64889f99e950f8b9e282070c9abb4305f1e0c59e99856efbc055"} Nov 25 17:16:00 crc kubenswrapper[4802]: I1125 17:16:00.154682 4802 generic.go:334] "Generic (PLEG): container finished" podID="eeca972a-357e-4fa8-a8b5-1421b6601219" containerID="6e8574a765f114ddbf8eaf165152b9db86fb7075e1afff0d5d953615c6548538" exitCode=1 Nov 25 17:16:00 crc kubenswrapper[4802]: I1125 17:16:00.154737 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" event={"ID":"eeca972a-357e-4fa8-a8b5-1421b6601219","Type":"ContainerDied","Data":"6e8574a765f114ddbf8eaf165152b9db86fb7075e1afff0d5d953615c6548538"} Nov 25 17:16:00 crc kubenswrapper[4802]: I1125 17:16:00.154784 4802 scope.go:117] "RemoveContainer" containerID="85bb3095155adb4276be27c6191c3475eb3bba0054184eac5e1cb2669bd15a66" Nov 25 17:16:00 crc kubenswrapper[4802]: I1125 17:16:00.155572 4802 scope.go:117] "RemoveContainer" containerID="6e8574a765f114ddbf8eaf165152b9db86fb7075e1afff0d5d953615c6548538" Nov 25 17:16:00 crc kubenswrapper[4802]: E1125 17:16:00.155853 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=glance-operator-controller-manager-6bfbf5c856-xw6fj_openstack-operators(eeca972a-357e-4fa8-a8b5-1421b6601219)\"" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" Nov 25 17:16:00 crc kubenswrapper[4802]: I1125 17:16:00.505100 4802 scope.go:117] "RemoveContainer" containerID="1dddde497d38cccc79697be659e98a97c142049481be6f0e11bc8312d99da3af" Nov 25 17:16:01 crc kubenswrapper[4802]: I1125 17:16:01.167835 4802 generic.go:334] "Generic (PLEG): container finished" podID="a1724655-9ac1-43dc-8292-f39870b4a855" containerID="f8f0e345d33296f54e2a57a4b164cab5f5ab1e0dce7cb3c77eb2564ed77b9f40" exitCode=1 Nov 25 17:16:01 crc kubenswrapper[4802]: I1125 17:16:01.168017 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" event={"ID":"a1724655-9ac1-43dc-8292-f39870b4a855","Type":"ContainerDied","Data":"f8f0e345d33296f54e2a57a4b164cab5f5ab1e0dce7cb3c77eb2564ed77b9f40"} Nov 25 17:16:01 crc kubenswrapper[4802]: I1125 17:16:01.168578 4802 scope.go:117] "RemoveContainer" containerID="1dddde497d38cccc79697be659e98a97c142049481be6f0e11bc8312d99da3af" Nov 25 17:16:01 crc kubenswrapper[4802]: I1125 17:16:01.169453 4802 scope.go:117] "RemoveContainer" containerID="f8f0e345d33296f54e2a57a4b164cab5f5ab1e0dce7cb3c77eb2564ed77b9f40" Nov 25 17:16:01 crc kubenswrapper[4802]: E1125 17:16:01.170002 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=metallb-operator-controller-manager-548f76d7c-s5g2f_metallb-system(a1724655-9ac1-43dc-8292-f39870b4a855)\"" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" Nov 25 17:16:01 crc kubenswrapper[4802]: I1125 17:16:01.171666 4802 generic.go:334] "Generic (PLEG): container finished" podID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" containerID="a2251ba467ce64889f99e950f8b9e282070c9abb4305f1e0c59e99856efbc055" exitCode=1 Nov 25 17:16:01 crc kubenswrapper[4802]: I1125 17:16:01.171708 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" event={"ID":"373fd8b2-c469-46b8-b20e-8ecd875c4a39","Type":"ContainerDied","Data":"a2251ba467ce64889f99e950f8b9e282070c9abb4305f1e0c59e99856efbc055"} Nov 25 17:16:01 crc kubenswrapper[4802]: I1125 17:16:01.172031 4802 scope.go:117] "RemoveContainer" containerID="a2251ba467ce64889f99e950f8b9e282070c9abb4305f1e0c59e99856efbc055" Nov 25 17:16:01 crc kubenswrapper[4802]: E1125 17:16:01.172293 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=operator pod=rabbitmq-cluster-operator-779fc9694b-987nc_openstack-operators(373fd8b2-c469-46b8-b20e-8ecd875c4a39)\"" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" Nov 25 17:16:01 crc kubenswrapper[4802]: I1125 17:16:01.218560 4802 scope.go:117] "RemoveContainer" containerID="f59fe8a229e461bdaeef0cc7495dc1200a263f9210f6e3aaf0c2a051ac50c485" Nov 25 17:16:01 crc kubenswrapper[4802]: I1125 17:16:01.504493 4802 scope.go:117] "RemoveContainer" containerID="b937b9f3a5d081aa5c27c6bff3a0ac96523739001944f058e149959f1d8068dc" Nov 25 17:16:01 crc kubenswrapper[4802]: I1125 17:16:01.710873 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 17:16:02 crc kubenswrapper[4802]: I1125 17:16:02.186842 4802 generic.go:334] "Generic (PLEG): container finished" podID="a5bdabd0-7540-4d45-8675-9bd88c415957" containerID="52ee74c874912a081d3846daf1b894f9235b49b48211eff209d228f65b1ea8f3" exitCode=1 Nov 25 17:16:02 crc kubenswrapper[4802]: I1125 17:16:02.186887 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" event={"ID":"a5bdabd0-7540-4d45-8675-9bd88c415957","Type":"ContainerDied","Data":"52ee74c874912a081d3846daf1b894f9235b49b48211eff209d228f65b1ea8f3"} Nov 25 17:16:02 crc kubenswrapper[4802]: I1125 17:16:02.186918 4802 scope.go:117] "RemoveContainer" containerID="b937b9f3a5d081aa5c27c6bff3a0ac96523739001944f058e149959f1d8068dc" Nov 25 17:16:02 crc kubenswrapper[4802]: I1125 17:16:02.187661 4802 scope.go:117] "RemoveContainer" containerID="52ee74c874912a081d3846daf1b894f9235b49b48211eff209d228f65b1ea8f3" Nov 25 17:16:02 crc kubenswrapper[4802]: E1125 17:16:02.188115 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-65c6fdb94d-5cfbr_openstack-operators(a5bdabd0-7540-4d45-8675-9bd88c415957)\"" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" Nov 25 17:16:02 crc kubenswrapper[4802]: I1125 17:16:02.354028 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:16:02 crc kubenswrapper[4802]: I1125 17:16:02.354975 4802 scope.go:117] "RemoveContainer" containerID="e903f73446fc13ae8e0c35edd104f75f1b890874642288c1dc68bfe6345bc27e" Nov 25 17:16:03 crc kubenswrapper[4802]: I1125 17:16:03.196090 4802 generic.go:334] "Generic (PLEG): container finished" podID="c328497f-325f-4500-ba48-d1523cd7ee74" containerID="c737609b983233b2182d834eab62f10b35be99a9ca3e5a9e9db9a2921a36554a" exitCode=1 Nov 25 17:16:03 crc kubenswrapper[4802]: I1125 17:16:03.196181 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" event={"ID":"c328497f-325f-4500-ba48-d1523cd7ee74","Type":"ContainerDied","Data":"c737609b983233b2182d834eab62f10b35be99a9ca3e5a9e9db9a2921a36554a"} Nov 25 17:16:03 crc kubenswrapper[4802]: I1125 17:16:03.196393 4802 scope.go:117] "RemoveContainer" containerID="e903f73446fc13ae8e0c35edd104f75f1b890874642288c1dc68bfe6345bc27e" Nov 25 17:16:03 crc kubenswrapper[4802]: I1125 17:16:03.196955 4802 scope.go:117] "RemoveContainer" containerID="c737609b983233b2182d834eab62f10b35be99a9ca3e5a9e9db9a2921a36554a" Nov 25 17:16:03 crc kubenswrapper[4802]: E1125 17:16:03.197272 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-d58bc84f4-fjc7z_openstack-operators(c328497f-325f-4500-ba48-d1523cd7ee74)\"" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" Nov 25 17:16:03 crc kubenswrapper[4802]: I1125 17:16:03.504107 4802 scope.go:117] "RemoveContainer" containerID="9ce5ad58c88b39008d469294bae4d4e6c1de715142944b1387e62b498caf2b4f" Nov 25 17:16:03 crc kubenswrapper[4802]: I1125 17:16:03.775617 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 17:16:03 crc kubenswrapper[4802]: I1125 17:16:03.927979 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 17:16:04 crc kubenswrapper[4802]: I1125 17:16:04.077630 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 17:16:04 crc kubenswrapper[4802]: I1125 17:16:04.181598 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 17:16:04 crc kubenswrapper[4802]: I1125 17:16:04.208423 4802 generic.go:334] "Generic (PLEG): container finished" podID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" containerID="6d67eec28a7c2a35e321c5847fd5362dae8084139ca8e13528c2a3037924936b" exitCode=1 Nov 25 17:16:04 crc kubenswrapper[4802]: I1125 17:16:04.208516 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" event={"ID":"eda71d4e-8d95-4944-a0b8-a8ac84c01ca4","Type":"ContainerDied","Data":"6d67eec28a7c2a35e321c5847fd5362dae8084139ca8e13528c2a3037924936b"} Nov 25 17:16:04 crc kubenswrapper[4802]: I1125 17:16:04.209455 4802 scope.go:117] "RemoveContainer" containerID="9ce5ad58c88b39008d469294bae4d4e6c1de715142944b1387e62b498caf2b4f" Nov 25 17:16:04 crc kubenswrapper[4802]: I1125 17:16:04.210110 4802 scope.go:117] "RemoveContainer" containerID="6d67eec28a7c2a35e321c5847fd5362dae8084139ca8e13528c2a3037924936b" Nov 25 17:16:04 crc kubenswrapper[4802]: E1125 17:16:04.210452 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-6fccf5cdb-rrwcf_openstack-operators(eda71d4e-8d95-4944-a0b8-a8ac84c01ca4)\"" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" Nov 25 17:16:04 crc kubenswrapper[4802]: I1125 17:16:04.315709 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 17:16:04 crc kubenswrapper[4802]: I1125 17:16:04.368813 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 17:16:04 crc kubenswrapper[4802]: I1125 17:16:04.529425 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-4d458" Nov 25 17:16:04 crc kubenswrapper[4802]: I1125 17:16:04.827898 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Nov 25 17:16:05 crc kubenswrapper[4802]: I1125 17:16:05.055990 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 17:16:05 crc kubenswrapper[4802]: I1125 17:16:05.138737 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 17:16:05 crc kubenswrapper[4802]: I1125 17:16:05.275462 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Nov 25 17:16:05 crc kubenswrapper[4802]: I1125 17:16:05.312882 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 17:16:05 crc kubenswrapper[4802]: I1125 17:16:05.334492 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 17:16:05 crc kubenswrapper[4802]: I1125 17:16:05.471427 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 17:16:05 crc kubenswrapper[4802]: I1125 17:16:05.473800 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 17:16:05 crc kubenswrapper[4802]: I1125 17:16:05.571700 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 17:16:05 crc kubenswrapper[4802]: I1125 17:16:05.791647 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 17:16:05 crc kubenswrapper[4802]: I1125 17:16:05.846273 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:16:05 crc kubenswrapper[4802]: I1125 17:16:05.847057 4802 scope.go:117] "RemoveContainer" containerID="16316a73afcdc8c5e68f8dfba330f336394584b43d0a454011d7380e169e24f1" Nov 25 17:16:05 crc kubenswrapper[4802]: E1125 17:16:05.847476 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=mariadb-operator-controller-manager-5959d99654-4j2qc_openstack-operators(5004bc39-3a06-4842-9565-8501f35c8c4b)\"" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" podUID="5004bc39-3a06-4842-9565-8501f35c8c4b" Nov 25 17:16:05 crc kubenswrapper[4802]: I1125 17:16:05.934038 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 17:16:05 crc kubenswrapper[4802]: I1125 17:16:05.935193 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 17:16:05 crc kubenswrapper[4802]: I1125 17:16:05.966434 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-qmlbh" Nov 25 17:16:06 crc kubenswrapper[4802]: I1125 17:16:06.013138 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 17:16:06 crc kubenswrapper[4802]: I1125 17:16:06.145909 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-thlvn" Nov 25 17:16:06 crc kubenswrapper[4802]: I1125 17:16:06.310648 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" Nov 25 17:16:06 crc kubenswrapper[4802]: I1125 17:16:06.312954 4802 scope.go:117] "RemoveContainer" containerID="f8f0e345d33296f54e2a57a4b164cab5f5ab1e0dce7cb3c77eb2564ed77b9f40" Nov 25 17:16:06 crc kubenswrapper[4802]: E1125 17:16:06.313371 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=metallb-operator-controller-manager-548f76d7c-s5g2f_metallb-system(a1724655-9ac1-43dc-8292-f39870b4a855)\"" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" Nov 25 17:16:06 crc kubenswrapper[4802]: I1125 17:16:06.326490 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 17:16:06 crc kubenswrapper[4802]: I1125 17:16:06.480374 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Nov 25 17:16:06 crc kubenswrapper[4802]: I1125 17:16:06.480380 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 25 17:16:06 crc kubenswrapper[4802]: I1125 17:16:06.504774 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:16:06 crc kubenswrapper[4802]: E1125 17:16:06.505163 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:16:06 crc kubenswrapper[4802]: I1125 17:16:06.511312 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 17:16:06 crc kubenswrapper[4802]: I1125 17:16:06.544674 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 17:16:06 crc kubenswrapper[4802]: I1125 17:16:06.607738 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 17:16:06 crc kubenswrapper[4802]: I1125 17:16:06.655205 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 17:16:06 crc kubenswrapper[4802]: I1125 17:16:06.669450 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 17:16:06 crc kubenswrapper[4802]: I1125 17:16:06.697898 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 17:16:06 crc kubenswrapper[4802]: I1125 17:16:06.733399 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 25 17:16:06 crc kubenswrapper[4802]: I1125 17:16:06.741206 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 17:16:06 crc kubenswrapper[4802]: I1125 17:16:06.873944 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Nov 25 17:16:06 crc kubenswrapper[4802]: I1125 17:16:06.979419 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.083350 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-q2wcj" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.132410 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.146486 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.159546 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.195014 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.195100 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-8dk78" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.259046 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.324882 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.427524 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.430276 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.558479 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.577917 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.603548 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.747498 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.763026 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.792072 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.824499 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.865316 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.879510 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 17:16:07 crc kubenswrapper[4802]: I1125 17:16:07.966858 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.047450 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.062609 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.263228 4802 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.270723 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.270966 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.275517 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.285871 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=14.285857976 podStartE2EDuration="14.285857976s" podCreationTimestamp="2025-11-25 17:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:16:08.285115576 +0000 UTC m=+1751.429462762" watchObservedRunningTime="2025-11-25 17:16:08.285857976 +0000 UTC m=+1751.430205172" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.327472 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.369290 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-xrjdp" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.408659 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.408703 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.409295 4802 scope.go:117] "RemoveContainer" containerID="6d67eec28a7c2a35e321c5847fd5362dae8084139ca8e13528c2a3037924936b" Nov 25 17:16:08 crc kubenswrapper[4802]: E1125 17:16:08.409548 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-6fccf5cdb-rrwcf_openstack-operators(eda71d4e-8d95-4944-a0b8-a8ac84c01ca4)\"" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.504615 4802 scope.go:117] "RemoveContainer" containerID="93ea15cbbe4d99268f06277f2d93ce56c43989561de09ed1af2238675c896eac" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.516251 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.585194 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.699066 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.745215 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.775186 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.787566 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.794666 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.799312 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.870203 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.880955 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.935988 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.941799 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.971492 4802 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.981259 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 17:16:08 crc kubenswrapper[4802]: I1125 17:16:08.981971 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.046000 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.106221 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.106261 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.107240 4802 scope.go:117] "RemoveContainer" containerID="52ee74c874912a081d3846daf1b894f9235b49b48211eff209d228f65b1ea8f3" Nov 25 17:16:09 crc kubenswrapper[4802]: E1125 17:16:09.107623 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-65c6fdb94d-5cfbr_openstack-operators(a5bdabd0-7540-4d45-8675-9bd88c415957)\"" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.110111 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.164884 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.188764 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.199649 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.200882 4802 scope.go:117] "RemoveContainer" containerID="6e8574a765f114ddbf8eaf165152b9db86fb7075e1afff0d5d953615c6548538" Nov 25 17:16:09 crc kubenswrapper[4802]: E1125 17:16:09.201435 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=glance-operator-controller-manager-6bfbf5c856-xw6fj_openstack-operators(eeca972a-357e-4fa8-a8b5-1421b6601219)\"" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.269173 4802 generic.go:334] "Generic (PLEG): container finished" podID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" containerID="976855d27304382f3214e6e2c2d67cb7d7a49c3f7665ffa312c37fc17ba152f1" exitCode=1 Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.269764 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" event={"ID":"3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb","Type":"ContainerDied","Data":"976855d27304382f3214e6e2c2d67cb7d7a49c3f7665ffa312c37fc17ba152f1"} Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.269924 4802 scope.go:117] "RemoveContainer" containerID="93ea15cbbe4d99268f06277f2d93ce56c43989561de09ed1af2238675c896eac" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.270453 4802 scope.go:117] "RemoveContainer" containerID="976855d27304382f3214e6e2c2d67cb7d7a49c3f7665ffa312c37fc17ba152f1" Nov 25 17:16:09 crc kubenswrapper[4802]: E1125 17:16:09.270732 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-5787d846c-g6b4c_openstack-operators(3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb)\"" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.319944 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.330011 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.342762 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.356176 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.389896 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.447634 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.518930 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.570275 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.580440 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-xh9p6" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.718106 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-8pstl" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.875582 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.988999 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 17:16:09 crc kubenswrapper[4802]: I1125 17:16:09.994096 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.051343 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.086533 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.087266 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.112803 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.207835 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.208772 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-xss8s" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.228853 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-jfwbm" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.232164 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-8z52s" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.286623 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.295745 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-f9sj6" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.406036 4802 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.444630 4802 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.468625 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.530550 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.575538 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.583164 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.584717 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-dllpk" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.593772 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.681539 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-v2c2s" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.725586 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.750582 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.754646 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-9j9ft" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.780029 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.814563 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.964851 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.980552 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 17:16:10 crc kubenswrapper[4802]: I1125 17:16:10.984742 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.111292 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.162775 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.182308 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.192302 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-7zbkx" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.227034 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.229762 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.260767 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.305970 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.307286 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.363264 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.391348 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.424204 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.438783 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.446670 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.505730 4802 scope.go:117] "RemoveContainer" containerID="a2251ba467ce64889f99e950f8b9e282070c9abb4305f1e0c59e99856efbc055" Nov 25 17:16:11 crc kubenswrapper[4802]: E1125 17:16:11.506174 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=operator pod=rabbitmq-cluster-operator-779fc9694b-987nc_openstack-operators(373fd8b2-c469-46b8-b20e-8ecd875c4a39)\"" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" podUID="373fd8b2-c469-46b8-b20e-8ecd875c4a39" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.515018 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.563435 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.583436 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.613289 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.645921 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-xcg2k" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.685537 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.754786 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.797841 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.825085 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.890112 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.914058 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.915148 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.930572 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 25 17:16:11 crc kubenswrapper[4802]: I1125 17:16:11.960401 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 17:16:12 crc kubenswrapper[4802]: I1125 17:16:12.007906 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 17:16:12 crc kubenswrapper[4802]: I1125 17:16:12.012364 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 17:16:12 crc kubenswrapper[4802]: I1125 17:16:12.093600 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 17:16:12 crc kubenswrapper[4802]: I1125 17:16:12.119219 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 17:16:12 crc kubenswrapper[4802]: I1125 17:16:12.126543 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 17:16:12 crc kubenswrapper[4802]: I1125 17:16:12.128582 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 17:16:12 crc kubenswrapper[4802]: I1125 17:16:12.303995 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 17:16:12 crc kubenswrapper[4802]: I1125 17:16:12.354416 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:16:12 crc kubenswrapper[4802]: I1125 17:16:12.355143 4802 scope.go:117] "RemoveContainer" containerID="c737609b983233b2182d834eab62f10b35be99a9ca3e5a9e9db9a2921a36554a" Nov 25 17:16:12 crc kubenswrapper[4802]: E1125 17:16:12.355473 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-d58bc84f4-fjc7z_openstack-operators(c328497f-325f-4500-ba48-d1523cd7ee74)\"" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" Nov 25 17:16:12 crc kubenswrapper[4802]: I1125 17:16:12.681586 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 17:16:12 crc kubenswrapper[4802]: I1125 17:16:12.700606 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 17:16:12 crc kubenswrapper[4802]: I1125 17:16:12.741932 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Nov 25 17:16:12 crc kubenswrapper[4802]: I1125 17:16:12.906988 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 17:16:12 crc kubenswrapper[4802]: I1125 17:16:12.910948 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 17:16:12 crc kubenswrapper[4802]: I1125 17:16:12.967419 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.082170 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.094215 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.125794 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.137814 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.291208 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.396282 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-rq8mq" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.506343 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.522788 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.538754 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.619821 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.644792 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.652840 4802 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.700873 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.704108 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.749275 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.797644 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.831404 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.885038 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.903006 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.937889 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.954326 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 17:16:13 crc kubenswrapper[4802]: I1125 17:16:13.971239 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.161899 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.163220 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.180411 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.207559 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.210058 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.224239 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-xdw54" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.250587 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.267010 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.422426 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.450809 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.522441 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.619630 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.699224 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.712968 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.773872 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.775958 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.784603 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.827567 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.852845 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-rxfjg" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.921591 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.922447 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.941775 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 17:16:14 crc kubenswrapper[4802]: I1125 17:16:14.991689 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.012974 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.026622 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.048334 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.060674 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.101924 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.111108 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.117392 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.180770 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.290409 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.290491 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.290935 4802 scope.go:117] "RemoveContainer" containerID="976855d27304382f3214e6e2c2d67cb7d7a49c3f7665ffa312c37fc17ba152f1" Nov 25 17:16:15 crc kubenswrapper[4802]: E1125 17:16:15.291272 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-5787d846c-g6b4c_openstack-operators(3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb)\"" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.313735 4802 scope.go:117] "RemoveContainer" containerID="976855d27304382f3214e6e2c2d67cb7d7a49c3f7665ffa312c37fc17ba152f1" Nov 25 17:16:15 crc kubenswrapper[4802]: E1125 17:16:15.314080 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-5787d846c-g6b4c_openstack-operators(3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb)\"" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" podUID="3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.327424 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.338352 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.340075 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.363282 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.364875 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.370683 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.388034 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.400893 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.426928 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.452159 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.598702 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.627061 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.629731 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.636006 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.769810 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.831368 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.856585 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.904192 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.958805 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.972933 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.985360 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 17:16:15 crc kubenswrapper[4802]: I1125 17:16:15.993823 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.049603 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.139438 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.155476 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.222432 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.225791 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.283594 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.286787 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.289995 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.310041 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.330188 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.421395 4802 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.421602 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://199b60b60e389b91cf44d85dd7d0a0f78d6af307f5090038b9d637fe177c2f15" gracePeriod=5 Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.532929 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.619723 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.623756 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.626327 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.664330 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.674428 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.706609 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.717975 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.750215 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.800264 4802 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.808117 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 17:16:16 crc kubenswrapper[4802]: I1125 17:16:16.845581 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 17:16:17 crc kubenswrapper[4802]: I1125 17:16:17.046475 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 17:16:17 crc kubenswrapper[4802]: I1125 17:16:17.058343 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 17:16:17 crc kubenswrapper[4802]: I1125 17:16:17.069152 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-nlkzd" Nov 25 17:16:17 crc kubenswrapper[4802]: I1125 17:16:17.102735 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 17:16:17 crc kubenswrapper[4802]: I1125 17:16:17.204381 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 17:16:17 crc kubenswrapper[4802]: I1125 17:16:17.328797 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 17:16:17 crc kubenswrapper[4802]: I1125 17:16:17.410817 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 17:16:17 crc kubenswrapper[4802]: I1125 17:16:17.459444 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 17:16:17 crc kubenswrapper[4802]: I1125 17:16:17.520388 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 17:16:17 crc kubenswrapper[4802]: I1125 17:16:17.573988 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 17:16:17 crc kubenswrapper[4802]: I1125 17:16:17.582211 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-87pxg" Nov 25 17:16:17 crc kubenswrapper[4802]: I1125 17:16:17.600012 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 17:16:17 crc kubenswrapper[4802]: I1125 17:16:17.832448 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 17:16:17 crc kubenswrapper[4802]: I1125 17:16:17.832693 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 17:16:17 crc kubenswrapper[4802]: I1125 17:16:17.944790 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-s9mkz" Nov 25 17:16:17 crc kubenswrapper[4802]: I1125 17:16:17.945520 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 17:16:17 crc kubenswrapper[4802]: I1125 17:16:17.996807 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.040632 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.104001 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.153819 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.154699 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.177375 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.201686 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.243774 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.261781 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.265298 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.287551 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.360425 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.400941 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.409518 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.460135 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.581038 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.720521 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.794632 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.796321 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.909755 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 17:16:18 crc kubenswrapper[4802]: I1125 17:16:18.994508 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.016555 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.065729 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.176087 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.199769 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.200808 4802 scope.go:117] "RemoveContainer" containerID="6e8574a765f114ddbf8eaf165152b9db86fb7075e1afff0d5d953615c6548538" Nov 25 17:16:19 crc kubenswrapper[4802]: E1125 17:16:19.201153 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=glance-operator-controller-manager-6bfbf5c856-xw6fj_openstack-operators(eeca972a-357e-4fa8-a8b5-1421b6601219)\"" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" podUID="eeca972a-357e-4fa8-a8b5-1421b6601219" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.227402 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.288766 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.304218 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.321427 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.326243 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.466725 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.504643 4802 scope.go:117] "RemoveContainer" containerID="16316a73afcdc8c5e68f8dfba330f336394584b43d0a454011d7380e169e24f1" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.504949 4802 scope.go:117] "RemoveContainer" containerID="f8f0e345d33296f54e2a57a4b164cab5f5ab1e0dce7cb3c77eb2564ed77b9f40" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.505105 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:16:19 crc kubenswrapper[4802]: E1125 17:16:19.505193 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=metallb-operator-controller-manager-548f76d7c-s5g2f_metallb-system(a1724655-9ac1-43dc-8292-f39870b4a855)\"" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" podUID="a1724655-9ac1-43dc-8292-f39870b4a855" Nov 25 17:16:19 crc kubenswrapper[4802]: E1125 17:16:19.505310 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.525653 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.771262 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.777502 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.906398 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Nov 25 17:16:19 crc kubenswrapper[4802]: I1125 17:16:19.929322 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-mcsgb" Nov 25 17:16:20 crc kubenswrapper[4802]: I1125 17:16:20.052685 4802 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 25 17:16:20 crc kubenswrapper[4802]: I1125 17:16:20.274510 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 17:16:20 crc kubenswrapper[4802]: I1125 17:16:20.505601 4802 scope.go:117] "RemoveContainer" containerID="6d67eec28a7c2a35e321c5847fd5362dae8084139ca8e13528c2a3037924936b" Nov 25 17:16:20 crc kubenswrapper[4802]: E1125 17:16:20.506215 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-6fccf5cdb-rrwcf_openstack-operators(eda71d4e-8d95-4944-a0b8-a8ac84c01ca4)\"" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" podUID="eda71d4e-8d95-4944-a0b8-a8ac84c01ca4" Nov 25 17:16:20 crc kubenswrapper[4802]: I1125 17:16:20.752153 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 25 17:16:20 crc kubenswrapper[4802]: I1125 17:16:20.826233 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" event={"ID":"5004bc39-3a06-4842-9565-8501f35c8c4b","Type":"ContainerStarted","Data":"f58a4826e617d1489929ad95663f96470d7578dfd1f810bfa620db8183a17865"} Nov 25 17:16:20 crc kubenswrapper[4802]: I1125 17:16:20.826498 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:16:20 crc kubenswrapper[4802]: I1125 17:16:20.970168 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 17:16:21 crc kubenswrapper[4802]: I1125 17:16:21.483728 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 17:16:21 crc kubenswrapper[4802]: I1125 17:16:21.487718 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 17:16:21 crc kubenswrapper[4802]: I1125 17:16:21.505029 4802 scope.go:117] "RemoveContainer" containerID="52ee74c874912a081d3846daf1b894f9235b49b48211eff209d228f65b1ea8f3" Nov 25 17:16:21 crc kubenswrapper[4802]: E1125 17:16:21.505352 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-65c6fdb94d-5cfbr_openstack-operators(a5bdabd0-7540-4d45-8675-9bd88c415957)\"" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" podUID="a5bdabd0-7540-4d45-8675-9bd88c415957" Nov 25 17:16:21 crc kubenswrapper[4802]: I1125 17:16:21.730068 4802 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 17:16:21 crc kubenswrapper[4802]: I1125 17:16:21.838152 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 17:16:21 crc kubenswrapper[4802]: I1125 17:16:21.838202 4802 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="199b60b60e389b91cf44d85dd7d0a0f78d6af307f5090038b9d637fe177c2f15" exitCode=137 Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.029334 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.029401 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.123764 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.123860 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.123976 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.124057 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.124072 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.124097 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.124141 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.124202 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.124382 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.124986 4802 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.125025 4802 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.125057 4802 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.125074 4802 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.138539 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.226551 4802 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.354763 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.355526 4802 scope.go:117] "RemoveContainer" containerID="c737609b983233b2182d834eab62f10b35be99a9ca3e5a9e9db9a2921a36554a" Nov 25 17:16:22 crc kubenswrapper[4802]: E1125 17:16:22.355811 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-d58bc84f4-fjc7z_openstack-operators(c328497f-325f-4500-ba48-d1523cd7ee74)\"" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" podUID="c328497f-325f-4500-ba48-d1523cd7ee74" Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.847411 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.847521 4802 scope.go:117] "RemoveContainer" containerID="199b60b60e389b91cf44d85dd7d0a0f78d6af307f5090038b9d637fe177c2f15" Nov 25 17:16:22 crc kubenswrapper[4802]: I1125 17:16:22.847664 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 17:16:23 crc kubenswrapper[4802]: I1125 17:16:23.511990 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 25 17:16:24 crc kubenswrapper[4802]: I1125 17:16:24.504525 4802 scope.go:117] "RemoveContainer" containerID="a2251ba467ce64889f99e950f8b9e282070c9abb4305f1e0c59e99856efbc055" Nov 25 17:16:24 crc kubenswrapper[4802]: I1125 17:16:24.867291 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-987nc" event={"ID":"373fd8b2-c469-46b8-b20e-8ecd875c4a39","Type":"ContainerStarted","Data":"89b785bbd18de590bf46177c9845685a0dcba066f976e1d43d2a7af901bccab6"} Nov 25 17:16:25 crc kubenswrapper[4802]: I1125 17:16:25.850443 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5959d99654-4j2qc" Nov 25 17:16:29 crc kubenswrapper[4802]: I1125 17:16:29.504743 4802 scope.go:117] "RemoveContainer" containerID="976855d27304382f3214e6e2c2d67cb7d7a49c3f7665ffa312c37fc17ba152f1" Nov 25 17:16:29 crc kubenswrapper[4802]: I1125 17:16:29.928773 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" event={"ID":"3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb","Type":"ContainerStarted","Data":"d41b86aade8f6340fd05dbc90c078d505c6c090a25f40c4939a9499046396807"} Nov 25 17:16:29 crc kubenswrapper[4802]: I1125 17:16:29.929670 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:16:30 crc kubenswrapper[4802]: I1125 17:16:30.505184 4802 scope.go:117] "RemoveContainer" containerID="f8f0e345d33296f54e2a57a4b164cab5f5ab1e0dce7cb3c77eb2564ed77b9f40" Nov 25 17:16:30 crc kubenswrapper[4802]: I1125 17:16:30.941070 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" event={"ID":"a1724655-9ac1-43dc-8292-f39870b4a855","Type":"ContainerStarted","Data":"048eb4feeab8601f391fc90c593b2db407b7216222c6e637cf2b8c91f9021ffd"} Nov 25 17:16:30 crc kubenswrapper[4802]: I1125 17:16:30.941738 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" Nov 25 17:16:31 crc kubenswrapper[4802]: I1125 17:16:31.505741 4802 scope.go:117] "RemoveContainer" containerID="6e8574a765f114ddbf8eaf165152b9db86fb7075e1afff0d5d953615c6548538" Nov 25 17:16:31 crc kubenswrapper[4802]: I1125 17:16:31.951337 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" event={"ID":"eeca972a-357e-4fa8-a8b5-1421b6601219","Type":"ContainerStarted","Data":"4aaff130e81e53f53f92baa67475544f3ee890fff24b16b5cd40d1c1fb8d2ae7"} Nov 25 17:16:31 crc kubenswrapper[4802]: I1125 17:16:31.951680 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:16:32 crc kubenswrapper[4802]: I1125 17:16:32.504969 4802 scope.go:117] "RemoveContainer" containerID="6d67eec28a7c2a35e321c5847fd5362dae8084139ca8e13528c2a3037924936b" Nov 25 17:16:32 crc kubenswrapper[4802]: I1125 17:16:32.505344 4802 scope.go:117] "RemoveContainer" containerID="52ee74c874912a081d3846daf1b894f9235b49b48211eff209d228f65b1ea8f3" Nov 25 17:16:32 crc kubenswrapper[4802]: I1125 17:16:32.961631 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" event={"ID":"a5bdabd0-7540-4d45-8675-9bd88c415957","Type":"ContainerStarted","Data":"cd049ddf264796cd8d08930e636cf3a31cd2427f6f231f701333a5a707c6a0df"} Nov 25 17:16:32 crc kubenswrapper[4802]: I1125 17:16:32.961898 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:16:32 crc kubenswrapper[4802]: I1125 17:16:32.964711 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" event={"ID":"eda71d4e-8d95-4944-a0b8-a8ac84c01ca4","Type":"ContainerStarted","Data":"d222a08bc146624345e4af6ea32059ab1fed6abf0e5022f8b3a17b2a850f725e"} Nov 25 17:16:32 crc kubenswrapper[4802]: I1125 17:16:32.964941 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:16:33 crc kubenswrapper[4802]: I1125 17:16:33.505386 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:16:33 crc kubenswrapper[4802]: E1125 17:16:33.505712 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:16:35 crc kubenswrapper[4802]: I1125 17:16:35.297299 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5787d846c-g6b4c" Nov 25 17:16:36 crc kubenswrapper[4802]: I1125 17:16:36.506237 4802 scope.go:117] "RemoveContainer" containerID="c737609b983233b2182d834eab62f10b35be99a9ca3e5a9e9db9a2921a36554a" Nov 25 17:16:37 crc kubenswrapper[4802]: I1125 17:16:37.002366 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" event={"ID":"c328497f-325f-4500-ba48-d1523cd7ee74","Type":"ContainerStarted","Data":"3e0c67888d9afbfb4b5edbcb9f180b89739ff955cc0de448de9a56cf6f400936"} Nov 25 17:16:37 crc kubenswrapper[4802]: I1125 17:16:37.002833 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:16:38 crc kubenswrapper[4802]: I1125 17:16:38.413510 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6fccf5cdb-rrwcf" Nov 25 17:16:39 crc kubenswrapper[4802]: I1125 17:16:39.111085 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-65c6fdb94d-5cfbr" Nov 25 17:16:39 crc kubenswrapper[4802]: I1125 17:16:39.202854 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-6bfbf5c856-xw6fj" Nov 25 17:16:41 crc kubenswrapper[4802]: I1125 17:16:41.368852 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-47dzd"] Nov 25 17:16:41 crc kubenswrapper[4802]: E1125 17:16:41.369252 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 17:16:41 crc kubenswrapper[4802]: I1125 17:16:41.369277 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 17:16:41 crc kubenswrapper[4802]: E1125 17:16:41.369302 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" containerName="installer" Nov 25 17:16:41 crc kubenswrapper[4802]: I1125 17:16:41.369320 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" containerName="installer" Nov 25 17:16:41 crc kubenswrapper[4802]: I1125 17:16:41.369516 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a7f65e4-cae3-49e6-8f88-df9a7e115260" containerName="installer" Nov 25 17:16:41 crc kubenswrapper[4802]: I1125 17:16:41.369544 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 17:16:41 crc kubenswrapper[4802]: I1125 17:16:41.370711 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47dzd" Nov 25 17:16:41 crc kubenswrapper[4802]: I1125 17:16:41.385393 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-47dzd"] Nov 25 17:16:41 crc kubenswrapper[4802]: I1125 17:16:41.464091 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11227111-5bae-4453-a19e-33f65e1efbe2-catalog-content\") pod \"redhat-marketplace-47dzd\" (UID: \"11227111-5bae-4453-a19e-33f65e1efbe2\") " pod="openshift-marketplace/redhat-marketplace-47dzd" Nov 25 17:16:41 crc kubenswrapper[4802]: I1125 17:16:41.464282 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11227111-5bae-4453-a19e-33f65e1efbe2-utilities\") pod \"redhat-marketplace-47dzd\" (UID: \"11227111-5bae-4453-a19e-33f65e1efbe2\") " pod="openshift-marketplace/redhat-marketplace-47dzd" Nov 25 17:16:41 crc kubenswrapper[4802]: I1125 17:16:41.464373 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x8bf\" (UniqueName: \"kubernetes.io/projected/11227111-5bae-4453-a19e-33f65e1efbe2-kube-api-access-7x8bf\") pod \"redhat-marketplace-47dzd\" (UID: \"11227111-5bae-4453-a19e-33f65e1efbe2\") " pod="openshift-marketplace/redhat-marketplace-47dzd" Nov 25 17:16:41 crc kubenswrapper[4802]: I1125 17:16:41.568228 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x8bf\" (UniqueName: \"kubernetes.io/projected/11227111-5bae-4453-a19e-33f65e1efbe2-kube-api-access-7x8bf\") pod \"redhat-marketplace-47dzd\" (UID: \"11227111-5bae-4453-a19e-33f65e1efbe2\") " pod="openshift-marketplace/redhat-marketplace-47dzd" Nov 25 17:16:41 crc kubenswrapper[4802]: I1125 17:16:41.568795 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11227111-5bae-4453-a19e-33f65e1efbe2-catalog-content\") pod \"redhat-marketplace-47dzd\" (UID: \"11227111-5bae-4453-a19e-33f65e1efbe2\") " pod="openshift-marketplace/redhat-marketplace-47dzd" Nov 25 17:16:41 crc kubenswrapper[4802]: I1125 17:16:41.568882 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11227111-5bae-4453-a19e-33f65e1efbe2-utilities\") pod \"redhat-marketplace-47dzd\" (UID: \"11227111-5bae-4453-a19e-33f65e1efbe2\") " pod="openshift-marketplace/redhat-marketplace-47dzd" Nov 25 17:16:41 crc kubenswrapper[4802]: I1125 17:16:41.572027 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11227111-5bae-4453-a19e-33f65e1efbe2-catalog-content\") pod \"redhat-marketplace-47dzd\" (UID: \"11227111-5bae-4453-a19e-33f65e1efbe2\") " pod="openshift-marketplace/redhat-marketplace-47dzd" Nov 25 17:16:41 crc kubenswrapper[4802]: I1125 17:16:41.572315 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11227111-5bae-4453-a19e-33f65e1efbe2-utilities\") pod \"redhat-marketplace-47dzd\" (UID: \"11227111-5bae-4453-a19e-33f65e1efbe2\") " pod="openshift-marketplace/redhat-marketplace-47dzd" Nov 25 17:16:41 crc kubenswrapper[4802]: I1125 17:16:41.593115 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x8bf\" (UniqueName: \"kubernetes.io/projected/11227111-5bae-4453-a19e-33f65e1efbe2-kube-api-access-7x8bf\") pod \"redhat-marketplace-47dzd\" (UID: \"11227111-5bae-4453-a19e-33f65e1efbe2\") " pod="openshift-marketplace/redhat-marketplace-47dzd" Nov 25 17:16:41 crc kubenswrapper[4802]: I1125 17:16:41.694740 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47dzd" Nov 25 17:16:42 crc kubenswrapper[4802]: I1125 17:16:42.146154 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-47dzd"] Nov 25 17:16:42 crc kubenswrapper[4802]: W1125 17:16:42.155160 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11227111_5bae_4453_a19e_33f65e1efbe2.slice/crio-c123210cd8bebc5dd570f20befc5432a80f4dcc6564cace8c385fd23ae817d0a WatchSource:0}: Error finding container c123210cd8bebc5dd570f20befc5432a80f4dcc6564cace8c385fd23ae817d0a: Status 404 returned error can't find the container with id c123210cd8bebc5dd570f20befc5432a80f4dcc6564cace8c385fd23ae817d0a Nov 25 17:16:42 crc kubenswrapper[4802]: I1125 17:16:42.359798 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-d58bc84f4-fjc7z" Nov 25 17:16:43 crc kubenswrapper[4802]: I1125 17:16:43.052250 4802 generic.go:334] "Generic (PLEG): container finished" podID="11227111-5bae-4453-a19e-33f65e1efbe2" containerID="20079a041d9ad58ef4fa2ea39fbeb9e4dd697a165faed77b73406539bf256b10" exitCode=0 Nov 25 17:16:43 crc kubenswrapper[4802]: I1125 17:16:43.052296 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47dzd" event={"ID":"11227111-5bae-4453-a19e-33f65e1efbe2","Type":"ContainerDied","Data":"20079a041d9ad58ef4fa2ea39fbeb9e4dd697a165faed77b73406539bf256b10"} Nov 25 17:16:43 crc kubenswrapper[4802]: I1125 17:16:43.052332 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47dzd" event={"ID":"11227111-5bae-4453-a19e-33f65e1efbe2","Type":"ContainerStarted","Data":"c123210cd8bebc5dd570f20befc5432a80f4dcc6564cace8c385fd23ae817d0a"} Nov 25 17:16:43 crc kubenswrapper[4802]: I1125 17:16:43.055547 4802 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 17:16:44 crc kubenswrapper[4802]: I1125 17:16:44.061268 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47dzd" event={"ID":"11227111-5bae-4453-a19e-33f65e1efbe2","Type":"ContainerStarted","Data":"1f1f764d4861c9b5bfe31f6698e8c9edf6623422448081cedda470dcb8dbad2e"} Nov 25 17:16:45 crc kubenswrapper[4802]: I1125 17:16:45.069803 4802 generic.go:334] "Generic (PLEG): container finished" podID="11227111-5bae-4453-a19e-33f65e1efbe2" containerID="1f1f764d4861c9b5bfe31f6698e8c9edf6623422448081cedda470dcb8dbad2e" exitCode=0 Nov 25 17:16:45 crc kubenswrapper[4802]: I1125 17:16:45.069880 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47dzd" event={"ID":"11227111-5bae-4453-a19e-33f65e1efbe2","Type":"ContainerDied","Data":"1f1f764d4861c9b5bfe31f6698e8c9edf6623422448081cedda470dcb8dbad2e"} Nov 25 17:16:46 crc kubenswrapper[4802]: I1125 17:16:46.081903 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47dzd" event={"ID":"11227111-5bae-4453-a19e-33f65e1efbe2","Type":"ContainerStarted","Data":"34dbaadf0f40d3a512d6ace116e22c1ee3f509cb07ea5ef31b9190b08a066cc4"} Nov 25 17:16:46 crc kubenswrapper[4802]: I1125 17:16:46.504306 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:16:46 crc kubenswrapper[4802]: E1125 17:16:46.504534 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:16:51 crc kubenswrapper[4802]: I1125 17:16:51.696210 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-47dzd" Nov 25 17:16:51 crc kubenswrapper[4802]: I1125 17:16:51.696628 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-47dzd" Nov 25 17:16:51 crc kubenswrapper[4802]: I1125 17:16:51.742891 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-47dzd" Nov 25 17:16:51 crc kubenswrapper[4802]: I1125 17:16:51.762622 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-47dzd" podStartSLOduration=8.164853943 podStartE2EDuration="10.762602512s" podCreationTimestamp="2025-11-25 17:16:41 +0000 UTC" firstStartedPulling="2025-11-25 17:16:43.055110325 +0000 UTC m=+1786.199457551" lastFinishedPulling="2025-11-25 17:16:45.652858914 +0000 UTC m=+1788.797206120" observedRunningTime="2025-11-25 17:16:46.100448331 +0000 UTC m=+1789.244795527" watchObservedRunningTime="2025-11-25 17:16:51.762602512 +0000 UTC m=+1794.906949698" Nov 25 17:16:52 crc kubenswrapper[4802]: I1125 17:16:52.161299 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-47dzd" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.888598 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.901652 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.901776 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.936406 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.937312 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.939557 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.939765 4802 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-vs6wx" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.941652 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.942086 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.957725 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.972227 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.972268 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82c98e65-c5e6-483a-a508-ba36360a5ff1-scripts\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.972294 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2kv6\" (UniqueName: \"kubernetes.io/projected/82c98e65-c5e6-483a-a508-ba36360a5ff1-kube-api-access-j2kv6\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.972327 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.972361 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/82c98e65-c5e6-483a-a508-ba36360a5ff1-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.972377 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.972395 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-dev\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.972423 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.972458 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-sys\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.972869 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82c98e65-c5e6-483a-a508-ba36360a5ff1-config-data\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.972927 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82c98e65-c5e6-483a-a508-ba36360a5ff1-logs\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.972950 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-run\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.972991 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:54 crc kubenswrapper[4802]: I1125 17:16:54.973028 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.074604 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5f67491f-4728-432f-ae26-a82724aade76-openstack-config-secret\") pod \"openstackclient\" (UID: \"5f67491f-4728-432f-ae26-a82724aade76\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.074681 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/82c98e65-c5e6-483a-a508-ba36360a5ff1-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.074714 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.074751 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-dev\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.074799 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtw9q\" (UniqueName: \"kubernetes.io/projected/5f67491f-4728-432f-ae26-a82724aade76-kube-api-access-qtw9q\") pod \"openstackclient\" (UID: \"5f67491f-4728-432f-ae26-a82724aade76\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.074826 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.074864 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-sys\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.074906 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82c98e65-c5e6-483a-a508-ba36360a5ff1-config-data\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.074933 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82c98e65-c5e6-483a-a508-ba36360a5ff1-logs\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.074958 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5f67491f-4728-432f-ae26-a82724aade76-openstack-config\") pod \"openstackclient\" (UID: \"5f67491f-4728-432f-ae26-a82724aade76\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.074982 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-run\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.075008 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/5f67491f-4728-432f-ae26-a82724aade76-openstack-scripts\") pod \"openstackclient\" (UID: \"5f67491f-4728-432f-ae26-a82724aade76\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.075033 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.075059 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.075069 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/82c98e65-c5e6-483a-a508-ba36360a5ff1-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.075103 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.075156 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82c98e65-c5e6-483a-a508-ba36360a5ff1-scripts\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.075197 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2kv6\" (UniqueName: \"kubernetes.io/projected/82c98e65-c5e6-483a-a508-ba36360a5ff1-kube-api-access-j2kv6\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.075224 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.075656 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.075698 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-dev\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.075746 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.075775 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-sys\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.075814 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-run\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.075813 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.075849 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.075650 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.075935 4802 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.076465 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82c98e65-c5e6-483a-a508-ba36360a5ff1-logs\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.080287 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82c98e65-c5e6-483a-a508-ba36360a5ff1-scripts\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.080988 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82c98e65-c5e6-483a-a508-ba36360a5ff1-config-data\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.098137 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2kv6\" (UniqueName: \"kubernetes.io/projected/82c98e65-c5e6-483a-a508-ba36360a5ff1-kube-api-access-j2kv6\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.103622 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.106834 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.177067 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtw9q\" (UniqueName: \"kubernetes.io/projected/5f67491f-4728-432f-ae26-a82724aade76-kube-api-access-qtw9q\") pod \"openstackclient\" (UID: \"5f67491f-4728-432f-ae26-a82724aade76\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.177344 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5f67491f-4728-432f-ae26-a82724aade76-openstack-config\") pod \"openstackclient\" (UID: \"5f67491f-4728-432f-ae26-a82724aade76\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.177372 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/5f67491f-4728-432f-ae26-a82724aade76-openstack-scripts\") pod \"openstackclient\" (UID: \"5f67491f-4728-432f-ae26-a82724aade76\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.177454 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5f67491f-4728-432f-ae26-a82724aade76-openstack-config-secret\") pod \"openstackclient\" (UID: \"5f67491f-4728-432f-ae26-a82724aade76\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.178382 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/5f67491f-4728-432f-ae26-a82724aade76-openstack-scripts\") pod \"openstackclient\" (UID: \"5f67491f-4728-432f-ae26-a82724aade76\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.178502 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5f67491f-4728-432f-ae26-a82724aade76-openstack-config\") pod \"openstackclient\" (UID: \"5f67491f-4728-432f-ae26-a82724aade76\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.181179 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5f67491f-4728-432f-ae26-a82724aade76-openstack-config-secret\") pod \"openstackclient\" (UID: \"5f67491f-4728-432f-ae26-a82724aade76\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.195116 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtw9q\" (UniqueName: \"kubernetes.io/projected/5f67491f-4728-432f-ae26-a82724aade76-kube-api-access-qtw9q\") pod \"openstackclient\" (UID: \"5f67491f-4728-432f-ae26-a82724aade76\") " pod="glance-kuttl-tests/openstackclient" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.230696 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.265567 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.550442 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 17:16:55 crc kubenswrapper[4802]: I1125 17:16:55.695996 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 17:16:55 crc kubenswrapper[4802]: W1125 17:16:55.699891 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82c98e65_c5e6_483a_a508_ba36360a5ff1.slice/crio-f25aaa7da5481cc8e6454e9af04b8c230d99038a3aa4fbff10d2c628f4adc74e WatchSource:0}: Error finding container f25aaa7da5481cc8e6454e9af04b8c230d99038a3aa4fbff10d2c628f4adc74e: Status 404 returned error can't find the container with id f25aaa7da5481cc8e6454e9af04b8c230d99038a3aa4fbff10d2c628f4adc74e Nov 25 17:16:56 crc kubenswrapper[4802]: I1125 17:16:56.150276 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"5f67491f-4728-432f-ae26-a82724aade76","Type":"ContainerStarted","Data":"da8cb9ad844d8a865fdb949db280535e13d080454bd55de828dddd7b96558ebb"} Nov 25 17:16:56 crc kubenswrapper[4802]: I1125 17:16:56.150855 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"5f67491f-4728-432f-ae26-a82724aade76","Type":"ContainerStarted","Data":"096f7c23bdf13b48c3a395d95a1f394460b86af38564cfe4ec28107d6fe3a931"} Nov 25 17:16:56 crc kubenswrapper[4802]: I1125 17:16:56.152929 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"82c98e65-c5e6-483a-a508-ba36360a5ff1","Type":"ContainerStarted","Data":"9abaa4cffc0151fd5b5b5a0449390ac0ac0795fb6a9bb4fbc535f3584848a5f2"} Nov 25 17:16:56 crc kubenswrapper[4802]: I1125 17:16:56.152969 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"82c98e65-c5e6-483a-a508-ba36360a5ff1","Type":"ContainerStarted","Data":"edc8ae7d5b99e2e0e0496e99d6c2051e85d188fbdaf35a5bc0ecac5849c93f08"} Nov 25 17:16:56 crc kubenswrapper[4802]: I1125 17:16:56.152980 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"82c98e65-c5e6-483a-a508-ba36360a5ff1","Type":"ContainerStarted","Data":"f25aaa7da5481cc8e6454e9af04b8c230d99038a3aa4fbff10d2c628f4adc74e"} Nov 25 17:16:57 crc kubenswrapper[4802]: I1125 17:16:57.161973 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"82c98e65-c5e6-483a-a508-ba36360a5ff1","Type":"ContainerStarted","Data":"68aac35737ade4b501624db70503535aff2e798666206c03c83715a2b1ab2bd4"} Nov 25 17:16:57 crc kubenswrapper[4802]: I1125 17:16:57.201606 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=3.201574021 podStartE2EDuration="3.201574021s" podCreationTimestamp="2025-11-25 17:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:16:56.202241296 +0000 UTC m=+1799.346588502" watchObservedRunningTime="2025-11-25 17:16:57.201574021 +0000 UTC m=+1800.345921257" Nov 25 17:16:57 crc kubenswrapper[4802]: I1125 17:16:57.206025 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=3.206011791 podStartE2EDuration="3.206011791s" podCreationTimestamp="2025-11-25 17:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:16:57.195026163 +0000 UTC m=+1800.339373369" watchObservedRunningTime="2025-11-25 17:16:57.206011791 +0000 UTC m=+1800.350358987" Nov 25 17:16:59 crc kubenswrapper[4802]: I1125 17:16:59.504491 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:16:59 crc kubenswrapper[4802]: E1125 17:16:59.505076 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:16:59 crc kubenswrapper[4802]: I1125 17:16:59.954563 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-47dzd"] Nov 25 17:16:59 crc kubenswrapper[4802]: I1125 17:16:59.954785 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-47dzd" podUID="11227111-5bae-4453-a19e-33f65e1efbe2" containerName="registry-server" containerID="cri-o://34dbaadf0f40d3a512d6ace116e22c1ee3f509cb07ea5ef31b9190b08a066cc4" gracePeriod=2 Nov 25 17:17:00 crc kubenswrapper[4802]: I1125 17:17:00.190631 4802 generic.go:334] "Generic (PLEG): container finished" podID="11227111-5bae-4453-a19e-33f65e1efbe2" containerID="34dbaadf0f40d3a512d6ace116e22c1ee3f509cb07ea5ef31b9190b08a066cc4" exitCode=0 Nov 25 17:17:00 crc kubenswrapper[4802]: I1125 17:17:00.190996 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47dzd" event={"ID":"11227111-5bae-4453-a19e-33f65e1efbe2","Type":"ContainerDied","Data":"34dbaadf0f40d3a512d6ace116e22c1ee3f509cb07ea5ef31b9190b08a066cc4"} Nov 25 17:17:00 crc kubenswrapper[4802]: I1125 17:17:00.357186 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47dzd" Nov 25 17:17:00 crc kubenswrapper[4802]: I1125 17:17:00.459631 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11227111-5bae-4453-a19e-33f65e1efbe2-catalog-content\") pod \"11227111-5bae-4453-a19e-33f65e1efbe2\" (UID: \"11227111-5bae-4453-a19e-33f65e1efbe2\") " Nov 25 17:17:00 crc kubenswrapper[4802]: I1125 17:17:00.459781 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11227111-5bae-4453-a19e-33f65e1efbe2-utilities\") pod \"11227111-5bae-4453-a19e-33f65e1efbe2\" (UID: \"11227111-5bae-4453-a19e-33f65e1efbe2\") " Nov 25 17:17:00 crc kubenswrapper[4802]: I1125 17:17:00.459811 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7x8bf\" (UniqueName: \"kubernetes.io/projected/11227111-5bae-4453-a19e-33f65e1efbe2-kube-api-access-7x8bf\") pod \"11227111-5bae-4453-a19e-33f65e1efbe2\" (UID: \"11227111-5bae-4453-a19e-33f65e1efbe2\") " Nov 25 17:17:00 crc kubenswrapper[4802]: I1125 17:17:00.460969 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11227111-5bae-4453-a19e-33f65e1efbe2-utilities" (OuterVolumeSpecName: "utilities") pod "11227111-5bae-4453-a19e-33f65e1efbe2" (UID: "11227111-5bae-4453-a19e-33f65e1efbe2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:17:00 crc kubenswrapper[4802]: I1125 17:17:00.476430 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11227111-5bae-4453-a19e-33f65e1efbe2-kube-api-access-7x8bf" (OuterVolumeSpecName: "kube-api-access-7x8bf") pod "11227111-5bae-4453-a19e-33f65e1efbe2" (UID: "11227111-5bae-4453-a19e-33f65e1efbe2"). InnerVolumeSpecName "kube-api-access-7x8bf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:17:00 crc kubenswrapper[4802]: I1125 17:17:00.477807 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11227111-5bae-4453-a19e-33f65e1efbe2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "11227111-5bae-4453-a19e-33f65e1efbe2" (UID: "11227111-5bae-4453-a19e-33f65e1efbe2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:17:00 crc kubenswrapper[4802]: I1125 17:17:00.561560 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11227111-5bae-4453-a19e-33f65e1efbe2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 17:17:00 crc kubenswrapper[4802]: I1125 17:17:00.561590 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11227111-5bae-4453-a19e-33f65e1efbe2-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 17:17:00 crc kubenswrapper[4802]: I1125 17:17:00.561600 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7x8bf\" (UniqueName: \"kubernetes.io/projected/11227111-5bae-4453-a19e-33f65e1efbe2-kube-api-access-7x8bf\") on node \"crc\" DevicePath \"\"" Nov 25 17:17:01 crc kubenswrapper[4802]: I1125 17:17:01.202174 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47dzd" event={"ID":"11227111-5bae-4453-a19e-33f65e1efbe2","Type":"ContainerDied","Data":"c123210cd8bebc5dd570f20befc5432a80f4dcc6564cace8c385fd23ae817d0a"} Nov 25 17:17:01 crc kubenswrapper[4802]: I1125 17:17:01.202234 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47dzd" Nov 25 17:17:01 crc kubenswrapper[4802]: I1125 17:17:01.202496 4802 scope.go:117] "RemoveContainer" containerID="34dbaadf0f40d3a512d6ace116e22c1ee3f509cb07ea5ef31b9190b08a066cc4" Nov 25 17:17:01 crc kubenswrapper[4802]: I1125 17:17:01.253871 4802 scope.go:117] "RemoveContainer" containerID="1f1f764d4861c9b5bfe31f6698e8c9edf6623422448081cedda470dcb8dbad2e" Nov 25 17:17:01 crc kubenswrapper[4802]: I1125 17:17:01.262008 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-47dzd"] Nov 25 17:17:01 crc kubenswrapper[4802]: I1125 17:17:01.269392 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-47dzd"] Nov 25 17:17:01 crc kubenswrapper[4802]: I1125 17:17:01.276352 4802 scope.go:117] "RemoveContainer" containerID="20079a041d9ad58ef4fa2ea39fbeb9e4dd697a165faed77b73406539bf256b10" Nov 25 17:17:01 crc kubenswrapper[4802]: I1125 17:17:01.513085 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11227111-5bae-4453-a19e-33f65e1efbe2" path="/var/lib/kubelet/pods/11227111-5bae-4453-a19e-33f65e1efbe2/volumes" Nov 25 17:17:04 crc kubenswrapper[4802]: I1125 17:17:04.362412 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-649tv"] Nov 25 17:17:04 crc kubenswrapper[4802]: E1125 17:17:04.362919 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11227111-5bae-4453-a19e-33f65e1efbe2" containerName="extract-content" Nov 25 17:17:04 crc kubenswrapper[4802]: I1125 17:17:04.362931 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="11227111-5bae-4453-a19e-33f65e1efbe2" containerName="extract-content" Nov 25 17:17:04 crc kubenswrapper[4802]: E1125 17:17:04.362951 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11227111-5bae-4453-a19e-33f65e1efbe2" containerName="extract-utilities" Nov 25 17:17:04 crc kubenswrapper[4802]: I1125 17:17:04.362958 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="11227111-5bae-4453-a19e-33f65e1efbe2" containerName="extract-utilities" Nov 25 17:17:04 crc kubenswrapper[4802]: E1125 17:17:04.362965 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11227111-5bae-4453-a19e-33f65e1efbe2" containerName="registry-server" Nov 25 17:17:04 crc kubenswrapper[4802]: I1125 17:17:04.362971 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="11227111-5bae-4453-a19e-33f65e1efbe2" containerName="registry-server" Nov 25 17:17:04 crc kubenswrapper[4802]: I1125 17:17:04.363100 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="11227111-5bae-4453-a19e-33f65e1efbe2" containerName="registry-server" Nov 25 17:17:04 crc kubenswrapper[4802]: I1125 17:17:04.364363 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-649tv" Nov 25 17:17:04 crc kubenswrapper[4802]: I1125 17:17:04.370842 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-649tv"] Nov 25 17:17:04 crc kubenswrapper[4802]: I1125 17:17:04.424750 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8cce376-3073-4192-b060-5dadc051766a-catalog-content\") pod \"redhat-marketplace-649tv\" (UID: \"c8cce376-3073-4192-b060-5dadc051766a\") " pod="openshift-marketplace/redhat-marketplace-649tv" Nov 25 17:17:04 crc kubenswrapper[4802]: I1125 17:17:04.424822 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rrqz\" (UniqueName: \"kubernetes.io/projected/c8cce376-3073-4192-b060-5dadc051766a-kube-api-access-4rrqz\") pod \"redhat-marketplace-649tv\" (UID: \"c8cce376-3073-4192-b060-5dadc051766a\") " pod="openshift-marketplace/redhat-marketplace-649tv" Nov 25 17:17:04 crc kubenswrapper[4802]: I1125 17:17:04.425156 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8cce376-3073-4192-b060-5dadc051766a-utilities\") pod \"redhat-marketplace-649tv\" (UID: \"c8cce376-3073-4192-b060-5dadc051766a\") " pod="openshift-marketplace/redhat-marketplace-649tv" Nov 25 17:17:04 crc kubenswrapper[4802]: I1125 17:17:04.528265 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8cce376-3073-4192-b060-5dadc051766a-catalog-content\") pod \"redhat-marketplace-649tv\" (UID: \"c8cce376-3073-4192-b060-5dadc051766a\") " pod="openshift-marketplace/redhat-marketplace-649tv" Nov 25 17:17:04 crc kubenswrapper[4802]: I1125 17:17:04.528397 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rrqz\" (UniqueName: \"kubernetes.io/projected/c8cce376-3073-4192-b060-5dadc051766a-kube-api-access-4rrqz\") pod \"redhat-marketplace-649tv\" (UID: \"c8cce376-3073-4192-b060-5dadc051766a\") " pod="openshift-marketplace/redhat-marketplace-649tv" Nov 25 17:17:04 crc kubenswrapper[4802]: I1125 17:17:04.528505 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8cce376-3073-4192-b060-5dadc051766a-utilities\") pod \"redhat-marketplace-649tv\" (UID: \"c8cce376-3073-4192-b060-5dadc051766a\") " pod="openshift-marketplace/redhat-marketplace-649tv" Nov 25 17:17:04 crc kubenswrapper[4802]: I1125 17:17:04.529067 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8cce376-3073-4192-b060-5dadc051766a-catalog-content\") pod \"redhat-marketplace-649tv\" (UID: \"c8cce376-3073-4192-b060-5dadc051766a\") " pod="openshift-marketplace/redhat-marketplace-649tv" Nov 25 17:17:04 crc kubenswrapper[4802]: I1125 17:17:04.530659 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8cce376-3073-4192-b060-5dadc051766a-utilities\") pod \"redhat-marketplace-649tv\" (UID: \"c8cce376-3073-4192-b060-5dadc051766a\") " pod="openshift-marketplace/redhat-marketplace-649tv" Nov 25 17:17:04 crc kubenswrapper[4802]: I1125 17:17:04.553840 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rrqz\" (UniqueName: \"kubernetes.io/projected/c8cce376-3073-4192-b060-5dadc051766a-kube-api-access-4rrqz\") pod \"redhat-marketplace-649tv\" (UID: \"c8cce376-3073-4192-b060-5dadc051766a\") " pod="openshift-marketplace/redhat-marketplace-649tv" Nov 25 17:17:04 crc kubenswrapper[4802]: I1125 17:17:04.681582 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-649tv" Nov 25 17:17:05 crc kubenswrapper[4802]: I1125 17:17:05.139209 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-649tv"] Nov 25 17:17:05 crc kubenswrapper[4802]: W1125 17:17:05.147756 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8cce376_3073_4192_b060_5dadc051766a.slice/crio-b75aa995d48311c00e582149c7dc42155214b78b8a9fe88c722c4271de6ba580 WatchSource:0}: Error finding container b75aa995d48311c00e582149c7dc42155214b78b8a9fe88c722c4271de6ba580: Status 404 returned error can't find the container with id b75aa995d48311c00e582149c7dc42155214b78b8a9fe88c722c4271de6ba580 Nov 25 17:17:05 crc kubenswrapper[4802]: I1125 17:17:05.231017 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:17:05 crc kubenswrapper[4802]: I1125 17:17:05.231065 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:17:05 crc kubenswrapper[4802]: I1125 17:17:05.231081 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:17:05 crc kubenswrapper[4802]: I1125 17:17:05.235367 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-649tv" event={"ID":"c8cce376-3073-4192-b060-5dadc051766a","Type":"ContainerStarted","Data":"b75aa995d48311c00e582149c7dc42155214b78b8a9fe88c722c4271de6ba580"} Nov 25 17:17:05 crc kubenswrapper[4802]: I1125 17:17:05.282425 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:17:05 crc kubenswrapper[4802]: I1125 17:17:05.288099 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:17:05 crc kubenswrapper[4802]: I1125 17:17:05.302397 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:17:06 crc kubenswrapper[4802]: I1125 17:17:06.244199 4802 generic.go:334] "Generic (PLEG): container finished" podID="c8cce376-3073-4192-b060-5dadc051766a" containerID="dc84635203477ce7b1613a8ff8daa15a7f840dafa044be8e05162a1f1cb1bc0c" exitCode=0 Nov 25 17:17:06 crc kubenswrapper[4802]: I1125 17:17:06.245295 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-649tv" event={"ID":"c8cce376-3073-4192-b060-5dadc051766a","Type":"ContainerDied","Data":"dc84635203477ce7b1613a8ff8daa15a7f840dafa044be8e05162a1f1cb1bc0c"} Nov 25 17:17:06 crc kubenswrapper[4802]: I1125 17:17:06.245340 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:17:06 crc kubenswrapper[4802]: I1125 17:17:06.245651 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:17:06 crc kubenswrapper[4802]: I1125 17:17:06.245701 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:17:06 crc kubenswrapper[4802]: I1125 17:17:06.259800 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:17:06 crc kubenswrapper[4802]: I1125 17:17:06.262547 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:17:06 crc kubenswrapper[4802]: I1125 17:17:06.269011 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:17:06 crc kubenswrapper[4802]: I1125 17:17:06.312829 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-548f76d7c-s5g2f" Nov 25 17:17:07 crc kubenswrapper[4802]: I1125 17:17:07.252545 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-649tv" event={"ID":"c8cce376-3073-4192-b060-5dadc051766a","Type":"ContainerStarted","Data":"c1584998f2a38b0c604b0c04bc1dfcb59096d942d01df19b043c30b081576a85"} Nov 25 17:17:08 crc kubenswrapper[4802]: I1125 17:17:08.263582 4802 generic.go:334] "Generic (PLEG): container finished" podID="c8cce376-3073-4192-b060-5dadc051766a" containerID="c1584998f2a38b0c604b0c04bc1dfcb59096d942d01df19b043c30b081576a85" exitCode=0 Nov 25 17:17:08 crc kubenswrapper[4802]: I1125 17:17:08.263680 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-649tv" event={"ID":"c8cce376-3073-4192-b060-5dadc051766a","Type":"ContainerDied","Data":"c1584998f2a38b0c604b0c04bc1dfcb59096d942d01df19b043c30b081576a85"} Nov 25 17:17:09 crc kubenswrapper[4802]: I1125 17:17:09.276471 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-649tv" event={"ID":"c8cce376-3073-4192-b060-5dadc051766a","Type":"ContainerStarted","Data":"94bd8b32081263dfca647b4dd016421c99a730aef8fe8a37386099669bfad830"} Nov 25 17:17:09 crc kubenswrapper[4802]: I1125 17:17:09.299147 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-649tv" podStartSLOduration=2.734053592 podStartE2EDuration="5.299110566s" podCreationTimestamp="2025-11-25 17:17:04 +0000 UTC" firstStartedPulling="2025-11-25 17:17:06.246998786 +0000 UTC m=+1809.391345972" lastFinishedPulling="2025-11-25 17:17:08.81205575 +0000 UTC m=+1811.956402946" observedRunningTime="2025-11-25 17:17:09.294885631 +0000 UTC m=+1812.439232817" watchObservedRunningTime="2025-11-25 17:17:09.299110566 +0000 UTC m=+1812.443457762" Nov 25 17:17:11 crc kubenswrapper[4802]: I1125 17:17:11.505028 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:17:11 crc kubenswrapper[4802]: E1125 17:17:11.505572 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:17:14 crc kubenswrapper[4802]: I1125 17:17:14.682209 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-649tv" Nov 25 17:17:14 crc kubenswrapper[4802]: I1125 17:17:14.682466 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-649tv" Nov 25 17:17:14 crc kubenswrapper[4802]: I1125 17:17:14.725088 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-649tv" Nov 25 17:17:15 crc kubenswrapper[4802]: I1125 17:17:15.369334 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-649tv" Nov 25 17:17:21 crc kubenswrapper[4802]: I1125 17:17:21.856480 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:17:21 crc kubenswrapper[4802]: E1125 17:17:21.857546 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:17:22 crc kubenswrapper[4802]: I1125 17:17:22.962280 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-649tv"] Nov 25 17:17:22 crc kubenswrapper[4802]: I1125 17:17:22.963656 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-649tv" podUID="c8cce376-3073-4192-b060-5dadc051766a" containerName="registry-server" containerID="cri-o://94bd8b32081263dfca647b4dd016421c99a730aef8fe8a37386099669bfad830" gracePeriod=2 Nov 25 17:17:23 crc kubenswrapper[4802]: I1125 17:17:23.390919 4802 generic.go:334] "Generic (PLEG): container finished" podID="c8cce376-3073-4192-b060-5dadc051766a" containerID="94bd8b32081263dfca647b4dd016421c99a730aef8fe8a37386099669bfad830" exitCode=0 Nov 25 17:17:23 crc kubenswrapper[4802]: I1125 17:17:23.390975 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-649tv" event={"ID":"c8cce376-3073-4192-b060-5dadc051766a","Type":"ContainerDied","Data":"94bd8b32081263dfca647b4dd016421c99a730aef8fe8a37386099669bfad830"} Nov 25 17:17:23 crc kubenswrapper[4802]: I1125 17:17:23.391336 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-649tv" event={"ID":"c8cce376-3073-4192-b060-5dadc051766a","Type":"ContainerDied","Data":"b75aa995d48311c00e582149c7dc42155214b78b8a9fe88c722c4271de6ba580"} Nov 25 17:17:23 crc kubenswrapper[4802]: I1125 17:17:23.391355 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b75aa995d48311c00e582149c7dc42155214b78b8a9fe88c722c4271de6ba580" Nov 25 17:17:23 crc kubenswrapper[4802]: I1125 17:17:23.405091 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-649tv" Nov 25 17:17:23 crc kubenswrapper[4802]: I1125 17:17:23.583323 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8cce376-3073-4192-b060-5dadc051766a-catalog-content\") pod \"c8cce376-3073-4192-b060-5dadc051766a\" (UID: \"c8cce376-3073-4192-b060-5dadc051766a\") " Nov 25 17:17:23 crc kubenswrapper[4802]: I1125 17:17:23.583485 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rrqz\" (UniqueName: \"kubernetes.io/projected/c8cce376-3073-4192-b060-5dadc051766a-kube-api-access-4rrqz\") pod \"c8cce376-3073-4192-b060-5dadc051766a\" (UID: \"c8cce376-3073-4192-b060-5dadc051766a\") " Nov 25 17:17:23 crc kubenswrapper[4802]: I1125 17:17:23.583542 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8cce376-3073-4192-b060-5dadc051766a-utilities\") pod \"c8cce376-3073-4192-b060-5dadc051766a\" (UID: \"c8cce376-3073-4192-b060-5dadc051766a\") " Nov 25 17:17:23 crc kubenswrapper[4802]: I1125 17:17:23.584430 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8cce376-3073-4192-b060-5dadc051766a-utilities" (OuterVolumeSpecName: "utilities") pod "c8cce376-3073-4192-b060-5dadc051766a" (UID: "c8cce376-3073-4192-b060-5dadc051766a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:17:23 crc kubenswrapper[4802]: I1125 17:17:23.603656 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8cce376-3073-4192-b060-5dadc051766a-kube-api-access-4rrqz" (OuterVolumeSpecName: "kube-api-access-4rrqz") pod "c8cce376-3073-4192-b060-5dadc051766a" (UID: "c8cce376-3073-4192-b060-5dadc051766a"). InnerVolumeSpecName "kube-api-access-4rrqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:17:23 crc kubenswrapper[4802]: I1125 17:17:23.623084 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8cce376-3073-4192-b060-5dadc051766a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8cce376-3073-4192-b060-5dadc051766a" (UID: "c8cce376-3073-4192-b060-5dadc051766a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:17:23 crc kubenswrapper[4802]: I1125 17:17:23.685029 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rrqz\" (UniqueName: \"kubernetes.io/projected/c8cce376-3073-4192-b060-5dadc051766a-kube-api-access-4rrqz\") on node \"crc\" DevicePath \"\"" Nov 25 17:17:23 crc kubenswrapper[4802]: I1125 17:17:23.685071 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8cce376-3073-4192-b060-5dadc051766a-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 17:17:23 crc kubenswrapper[4802]: I1125 17:17:23.685080 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8cce376-3073-4192-b060-5dadc051766a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 17:17:24 crc kubenswrapper[4802]: I1125 17:17:24.397576 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-649tv" Nov 25 17:17:24 crc kubenswrapper[4802]: I1125 17:17:24.427373 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-649tv"] Nov 25 17:17:24 crc kubenswrapper[4802]: I1125 17:17:24.434619 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-649tv"] Nov 25 17:17:25 crc kubenswrapper[4802]: I1125 17:17:25.513715 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8cce376-3073-4192-b060-5dadc051766a" path="/var/lib/kubelet/pods/c8cce376-3073-4192-b060-5dadc051766a/volumes" Nov 25 17:17:26 crc kubenswrapper[4802]: I1125 17:17:26.971078 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5ggk9"] Nov 25 17:17:26 crc kubenswrapper[4802]: E1125 17:17:26.971639 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8cce376-3073-4192-b060-5dadc051766a" containerName="extract-content" Nov 25 17:17:26 crc kubenswrapper[4802]: I1125 17:17:26.971650 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8cce376-3073-4192-b060-5dadc051766a" containerName="extract-content" Nov 25 17:17:26 crc kubenswrapper[4802]: E1125 17:17:26.971678 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8cce376-3073-4192-b060-5dadc051766a" containerName="extract-utilities" Nov 25 17:17:26 crc kubenswrapper[4802]: I1125 17:17:26.971684 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8cce376-3073-4192-b060-5dadc051766a" containerName="extract-utilities" Nov 25 17:17:26 crc kubenswrapper[4802]: E1125 17:17:26.971697 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8cce376-3073-4192-b060-5dadc051766a" containerName="registry-server" Nov 25 17:17:26 crc kubenswrapper[4802]: I1125 17:17:26.971703 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8cce376-3073-4192-b060-5dadc051766a" containerName="registry-server" Nov 25 17:17:26 crc kubenswrapper[4802]: I1125 17:17:26.971840 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8cce376-3073-4192-b060-5dadc051766a" containerName="registry-server" Nov 25 17:17:26 crc kubenswrapper[4802]: I1125 17:17:26.972801 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5ggk9" Nov 25 17:17:26 crc kubenswrapper[4802]: I1125 17:17:26.981267 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5ggk9"] Nov 25 17:17:27 crc kubenswrapper[4802]: I1125 17:17:27.035758 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pkqq\" (UniqueName: \"kubernetes.io/projected/16385bba-8b5a-42be-9357-bb2439bf2dc4-kube-api-access-7pkqq\") pod \"redhat-marketplace-5ggk9\" (UID: \"16385bba-8b5a-42be-9357-bb2439bf2dc4\") " pod="openshift-marketplace/redhat-marketplace-5ggk9" Nov 25 17:17:27 crc kubenswrapper[4802]: I1125 17:17:27.035825 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16385bba-8b5a-42be-9357-bb2439bf2dc4-utilities\") pod \"redhat-marketplace-5ggk9\" (UID: \"16385bba-8b5a-42be-9357-bb2439bf2dc4\") " pod="openshift-marketplace/redhat-marketplace-5ggk9" Nov 25 17:17:27 crc kubenswrapper[4802]: I1125 17:17:27.035858 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16385bba-8b5a-42be-9357-bb2439bf2dc4-catalog-content\") pod \"redhat-marketplace-5ggk9\" (UID: \"16385bba-8b5a-42be-9357-bb2439bf2dc4\") " pod="openshift-marketplace/redhat-marketplace-5ggk9" Nov 25 17:17:27 crc kubenswrapper[4802]: I1125 17:17:27.136930 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16385bba-8b5a-42be-9357-bb2439bf2dc4-catalog-content\") pod \"redhat-marketplace-5ggk9\" (UID: \"16385bba-8b5a-42be-9357-bb2439bf2dc4\") " pod="openshift-marketplace/redhat-marketplace-5ggk9" Nov 25 17:17:27 crc kubenswrapper[4802]: I1125 17:17:27.137110 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pkqq\" (UniqueName: \"kubernetes.io/projected/16385bba-8b5a-42be-9357-bb2439bf2dc4-kube-api-access-7pkqq\") pod \"redhat-marketplace-5ggk9\" (UID: \"16385bba-8b5a-42be-9357-bb2439bf2dc4\") " pod="openshift-marketplace/redhat-marketplace-5ggk9" Nov 25 17:17:27 crc kubenswrapper[4802]: I1125 17:17:27.137166 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16385bba-8b5a-42be-9357-bb2439bf2dc4-utilities\") pod \"redhat-marketplace-5ggk9\" (UID: \"16385bba-8b5a-42be-9357-bb2439bf2dc4\") " pod="openshift-marketplace/redhat-marketplace-5ggk9" Nov 25 17:17:27 crc kubenswrapper[4802]: I1125 17:17:27.137489 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16385bba-8b5a-42be-9357-bb2439bf2dc4-catalog-content\") pod \"redhat-marketplace-5ggk9\" (UID: \"16385bba-8b5a-42be-9357-bb2439bf2dc4\") " pod="openshift-marketplace/redhat-marketplace-5ggk9" Nov 25 17:17:27 crc kubenswrapper[4802]: I1125 17:17:27.137534 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16385bba-8b5a-42be-9357-bb2439bf2dc4-utilities\") pod \"redhat-marketplace-5ggk9\" (UID: \"16385bba-8b5a-42be-9357-bb2439bf2dc4\") " pod="openshift-marketplace/redhat-marketplace-5ggk9" Nov 25 17:17:27 crc kubenswrapper[4802]: I1125 17:17:27.158166 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pkqq\" (UniqueName: \"kubernetes.io/projected/16385bba-8b5a-42be-9357-bb2439bf2dc4-kube-api-access-7pkqq\") pod \"redhat-marketplace-5ggk9\" (UID: \"16385bba-8b5a-42be-9357-bb2439bf2dc4\") " pod="openshift-marketplace/redhat-marketplace-5ggk9" Nov 25 17:17:27 crc kubenswrapper[4802]: I1125 17:17:27.287708 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5ggk9" Nov 25 17:17:27 crc kubenswrapper[4802]: I1125 17:17:27.745746 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5ggk9"] Nov 25 17:17:27 crc kubenswrapper[4802]: W1125 17:17:27.750606 4802 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16385bba_8b5a_42be_9357_bb2439bf2dc4.slice/crio-89e43d9fe2ffd42481d5d3255e1d0a0ab3480bb3bea361f1d663b44da206bd6e WatchSource:0}: Error finding container 89e43d9fe2ffd42481d5d3255e1d0a0ab3480bb3bea361f1d663b44da206bd6e: Status 404 returned error can't find the container with id 89e43d9fe2ffd42481d5d3255e1d0a0ab3480bb3bea361f1d663b44da206bd6e Nov 25 17:17:28 crc kubenswrapper[4802]: I1125 17:17:28.437163 4802 generic.go:334] "Generic (PLEG): container finished" podID="16385bba-8b5a-42be-9357-bb2439bf2dc4" containerID="0fbf753bf1ccfb2ca11e00dd96a16d23da33d3c62c0e30d3844dd20435e8d72a" exitCode=0 Nov 25 17:17:28 crc kubenswrapper[4802]: I1125 17:17:28.437215 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ggk9" event={"ID":"16385bba-8b5a-42be-9357-bb2439bf2dc4","Type":"ContainerDied","Data":"0fbf753bf1ccfb2ca11e00dd96a16d23da33d3c62c0e30d3844dd20435e8d72a"} Nov 25 17:17:28 crc kubenswrapper[4802]: I1125 17:17:28.442320 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ggk9" event={"ID":"16385bba-8b5a-42be-9357-bb2439bf2dc4","Type":"ContainerStarted","Data":"89e43d9fe2ffd42481d5d3255e1d0a0ab3480bb3bea361f1d663b44da206bd6e"} Nov 25 17:17:29 crc kubenswrapper[4802]: I1125 17:17:29.452242 4802 generic.go:334] "Generic (PLEG): container finished" podID="16385bba-8b5a-42be-9357-bb2439bf2dc4" containerID="cc74930300789d532e334cd0c0a0af87dfbb754914e2bb5fad01711d93dbcc49" exitCode=0 Nov 25 17:17:29 crc kubenswrapper[4802]: I1125 17:17:29.452317 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ggk9" event={"ID":"16385bba-8b5a-42be-9357-bb2439bf2dc4","Type":"ContainerDied","Data":"cc74930300789d532e334cd0c0a0af87dfbb754914e2bb5fad01711d93dbcc49"} Nov 25 17:17:30 crc kubenswrapper[4802]: I1125 17:17:30.463290 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ggk9" event={"ID":"16385bba-8b5a-42be-9357-bb2439bf2dc4","Type":"ContainerStarted","Data":"2338638ca5ed88971ede29bb6724d04e0a9176a87cfa1aa9cade74a7f9a6b094"} Nov 25 17:17:30 crc kubenswrapper[4802]: I1125 17:17:30.483425 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5ggk9" podStartSLOduration=3.073558637 podStartE2EDuration="4.483407106s" podCreationTimestamp="2025-11-25 17:17:26 +0000 UTC" firstStartedPulling="2025-11-25 17:17:28.439032139 +0000 UTC m=+1831.583379325" lastFinishedPulling="2025-11-25 17:17:29.848880608 +0000 UTC m=+1832.993227794" observedRunningTime="2025-11-25 17:17:30.481978737 +0000 UTC m=+1833.626325933" watchObservedRunningTime="2025-11-25 17:17:30.483407106 +0000 UTC m=+1833.627754292" Nov 25 17:17:32 crc kubenswrapper[4802]: I1125 17:17:32.504577 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:17:33 crc kubenswrapper[4802]: I1125 17:17:33.484094 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerStarted","Data":"0ba8f23d451c7d83e7568a15af490bcf0f8abaf83f184136081d26b773dddc2e"} Nov 25 17:17:37 crc kubenswrapper[4802]: I1125 17:17:37.288538 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5ggk9" Nov 25 17:17:37 crc kubenswrapper[4802]: I1125 17:17:37.290265 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5ggk9" Nov 25 17:17:37 crc kubenswrapper[4802]: I1125 17:17:37.333240 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5ggk9" Nov 25 17:17:37 crc kubenswrapper[4802]: I1125 17:17:37.566423 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5ggk9" Nov 25 17:17:44 crc kubenswrapper[4802]: I1125 17:17:44.160908 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5ggk9"] Nov 25 17:17:44 crc kubenswrapper[4802]: I1125 17:17:44.161786 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5ggk9" podUID="16385bba-8b5a-42be-9357-bb2439bf2dc4" containerName="registry-server" containerID="cri-o://2338638ca5ed88971ede29bb6724d04e0a9176a87cfa1aa9cade74a7f9a6b094" gracePeriod=2 Nov 25 17:17:44 crc kubenswrapper[4802]: I1125 17:17:44.580063 4802 generic.go:334] "Generic (PLEG): container finished" podID="16385bba-8b5a-42be-9357-bb2439bf2dc4" containerID="2338638ca5ed88971ede29bb6724d04e0a9176a87cfa1aa9cade74a7f9a6b094" exitCode=0 Nov 25 17:17:44 crc kubenswrapper[4802]: I1125 17:17:44.580170 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ggk9" event={"ID":"16385bba-8b5a-42be-9357-bb2439bf2dc4","Type":"ContainerDied","Data":"2338638ca5ed88971ede29bb6724d04e0a9176a87cfa1aa9cade74a7f9a6b094"} Nov 25 17:17:44 crc kubenswrapper[4802]: I1125 17:17:44.580412 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ggk9" event={"ID":"16385bba-8b5a-42be-9357-bb2439bf2dc4","Type":"ContainerDied","Data":"89e43d9fe2ffd42481d5d3255e1d0a0ab3480bb3bea361f1d663b44da206bd6e"} Nov 25 17:17:44 crc kubenswrapper[4802]: I1125 17:17:44.580429 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89e43d9fe2ffd42481d5d3255e1d0a0ab3480bb3bea361f1d663b44da206bd6e" Nov 25 17:17:44 crc kubenswrapper[4802]: I1125 17:17:44.633393 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5ggk9" Nov 25 17:17:44 crc kubenswrapper[4802]: I1125 17:17:44.825666 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pkqq\" (UniqueName: \"kubernetes.io/projected/16385bba-8b5a-42be-9357-bb2439bf2dc4-kube-api-access-7pkqq\") pod \"16385bba-8b5a-42be-9357-bb2439bf2dc4\" (UID: \"16385bba-8b5a-42be-9357-bb2439bf2dc4\") " Nov 25 17:17:44 crc kubenswrapper[4802]: I1125 17:17:44.825855 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16385bba-8b5a-42be-9357-bb2439bf2dc4-catalog-content\") pod \"16385bba-8b5a-42be-9357-bb2439bf2dc4\" (UID: \"16385bba-8b5a-42be-9357-bb2439bf2dc4\") " Nov 25 17:17:44 crc kubenswrapper[4802]: I1125 17:17:44.825931 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16385bba-8b5a-42be-9357-bb2439bf2dc4-utilities\") pod \"16385bba-8b5a-42be-9357-bb2439bf2dc4\" (UID: \"16385bba-8b5a-42be-9357-bb2439bf2dc4\") " Nov 25 17:17:44 crc kubenswrapper[4802]: I1125 17:17:44.826765 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16385bba-8b5a-42be-9357-bb2439bf2dc4-utilities" (OuterVolumeSpecName: "utilities") pod "16385bba-8b5a-42be-9357-bb2439bf2dc4" (UID: "16385bba-8b5a-42be-9357-bb2439bf2dc4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:17:44 crc kubenswrapper[4802]: I1125 17:17:44.832834 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16385bba-8b5a-42be-9357-bb2439bf2dc4-kube-api-access-7pkqq" (OuterVolumeSpecName: "kube-api-access-7pkqq") pod "16385bba-8b5a-42be-9357-bb2439bf2dc4" (UID: "16385bba-8b5a-42be-9357-bb2439bf2dc4"). InnerVolumeSpecName "kube-api-access-7pkqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:17:44 crc kubenswrapper[4802]: I1125 17:17:44.845022 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16385bba-8b5a-42be-9357-bb2439bf2dc4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "16385bba-8b5a-42be-9357-bb2439bf2dc4" (UID: "16385bba-8b5a-42be-9357-bb2439bf2dc4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:17:44 crc kubenswrapper[4802]: I1125 17:17:44.927501 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pkqq\" (UniqueName: \"kubernetes.io/projected/16385bba-8b5a-42be-9357-bb2439bf2dc4-kube-api-access-7pkqq\") on node \"crc\" DevicePath \"\"" Nov 25 17:17:44 crc kubenswrapper[4802]: I1125 17:17:44.927560 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16385bba-8b5a-42be-9357-bb2439bf2dc4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 17:17:44 crc kubenswrapper[4802]: I1125 17:17:44.927579 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16385bba-8b5a-42be-9357-bb2439bf2dc4-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 17:17:45 crc kubenswrapper[4802]: I1125 17:17:45.588267 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5ggk9" Nov 25 17:17:45 crc kubenswrapper[4802]: I1125 17:17:45.608105 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5ggk9"] Nov 25 17:17:45 crc kubenswrapper[4802]: I1125 17:17:45.614347 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5ggk9"] Nov 25 17:17:47 crc kubenswrapper[4802]: I1125 17:17:47.522419 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16385bba-8b5a-42be-9357-bb2439bf2dc4" path="/var/lib/kubelet/pods/16385bba-8b5a-42be-9357-bb2439bf2dc4/volumes" Nov 25 17:18:03 crc kubenswrapper[4802]: E1125 17:18:03.912601 4802 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.245:35406->38.102.83.245:40353: read tcp 38.102.83.245:35406->38.102.83.245:40353: read: connection reset by peer Nov 25 17:18:03 crc kubenswrapper[4802]: E1125 17:18:03.913377 4802 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.245:35406->38.102.83.245:40353: write tcp 38.102.83.245:35406->38.102.83.245:40353: write: broken pipe Nov 25 17:18:04 crc kubenswrapper[4802]: I1125 17:18:04.204203 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 17:18:04 crc kubenswrapper[4802]: I1125 17:18:04.204657 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="82c98e65-c5e6-483a-a508-ba36360a5ff1" containerName="glance-log" containerID="cri-o://edc8ae7d5b99e2e0e0496e99d6c2051e85d188fbdaf35a5bc0ecac5849c93f08" gracePeriod=30 Nov 25 17:18:04 crc kubenswrapper[4802]: I1125 17:18:04.204744 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="82c98e65-c5e6-483a-a508-ba36360a5ff1" containerName="glance-api" containerID="cri-o://68aac35737ade4b501624db70503535aff2e798666206c03c83715a2b1ab2bd4" gracePeriod=30 Nov 25 17:18:04 crc kubenswrapper[4802]: I1125 17:18:04.204732 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="82c98e65-c5e6-483a-a508-ba36360a5ff1" containerName="glance-httpd" containerID="cri-o://9abaa4cffc0151fd5b5b5a0449390ac0ac0795fb6a9bb4fbc535f3584848a5f2" gracePeriod=30 Nov 25 17:18:04 crc kubenswrapper[4802]: I1125 17:18:04.783200 4802 generic.go:334] "Generic (PLEG): container finished" podID="82c98e65-c5e6-483a-a508-ba36360a5ff1" containerID="68aac35737ade4b501624db70503535aff2e798666206c03c83715a2b1ab2bd4" exitCode=0 Nov 25 17:18:04 crc kubenswrapper[4802]: I1125 17:18:04.783544 4802 generic.go:334] "Generic (PLEG): container finished" podID="82c98e65-c5e6-483a-a508-ba36360a5ff1" containerID="9abaa4cffc0151fd5b5b5a0449390ac0ac0795fb6a9bb4fbc535f3584848a5f2" exitCode=0 Nov 25 17:18:04 crc kubenswrapper[4802]: I1125 17:18:04.783565 4802 generic.go:334] "Generic (PLEG): container finished" podID="82c98e65-c5e6-483a-a508-ba36360a5ff1" containerID="edc8ae7d5b99e2e0e0496e99d6c2051e85d188fbdaf35a5bc0ecac5849c93f08" exitCode=143 Nov 25 17:18:04 crc kubenswrapper[4802]: I1125 17:18:04.783397 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"82c98e65-c5e6-483a-a508-ba36360a5ff1","Type":"ContainerDied","Data":"68aac35737ade4b501624db70503535aff2e798666206c03c83715a2b1ab2bd4"} Nov 25 17:18:04 crc kubenswrapper[4802]: I1125 17:18:04.783612 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"82c98e65-c5e6-483a-a508-ba36360a5ff1","Type":"ContainerDied","Data":"9abaa4cffc0151fd5b5b5a0449390ac0ac0795fb6a9bb4fbc535f3584848a5f2"} Nov 25 17:18:04 crc kubenswrapper[4802]: I1125 17:18:04.783634 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"82c98e65-c5e6-483a-a508-ba36360a5ff1","Type":"ContainerDied","Data":"edc8ae7d5b99e2e0e0496e99d6c2051e85d188fbdaf35a5bc0ecac5849c93f08"} Nov 25 17:18:04 crc kubenswrapper[4802]: I1125 17:18:04.911233 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.061434 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-run\") pod \"82c98e65-c5e6-483a-a508-ba36360a5ff1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.061492 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82c98e65-c5e6-483a-a508-ba36360a5ff1-scripts\") pod \"82c98e65-c5e6-483a-a508-ba36360a5ff1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.061578 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82c98e65-c5e6-483a-a508-ba36360a5ff1-logs\") pod \"82c98e65-c5e6-483a-a508-ba36360a5ff1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.061599 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-lib-modules\") pod \"82c98e65-c5e6-483a-a508-ba36360a5ff1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.061624 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-etc-iscsi\") pod \"82c98e65-c5e6-483a-a508-ba36360a5ff1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.061657 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-sys\") pod \"82c98e65-c5e6-483a-a508-ba36360a5ff1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.061679 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2kv6\" (UniqueName: \"kubernetes.io/projected/82c98e65-c5e6-483a-a508-ba36360a5ff1-kube-api-access-j2kv6\") pod \"82c98e65-c5e6-483a-a508-ba36360a5ff1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.061703 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-etc-nvme\") pod \"82c98e65-c5e6-483a-a508-ba36360a5ff1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.061729 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"82c98e65-c5e6-483a-a508-ba36360a5ff1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.061586 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-run" (OuterVolumeSpecName: "run") pod "82c98e65-c5e6-483a-a508-ba36360a5ff1" (UID: "82c98e65-c5e6-483a-a508-ba36360a5ff1"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.061808 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "82c98e65-c5e6-483a-a508-ba36360a5ff1" (UID: "82c98e65-c5e6-483a-a508-ba36360a5ff1"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.061800 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "82c98e65-c5e6-483a-a508-ba36360a5ff1" (UID: "82c98e65-c5e6-483a-a508-ba36360a5ff1"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.061751 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-dev\") pod \"82c98e65-c5e6-483a-a508-ba36360a5ff1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.061926 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-var-locks-brick\") pod \"82c98e65-c5e6-483a-a508-ba36360a5ff1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.061931 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-dev" (OuterVolumeSpecName: "dev") pod "82c98e65-c5e6-483a-a508-ba36360a5ff1" (UID: "82c98e65-c5e6-483a-a508-ba36360a5ff1"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.061987 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "82c98e65-c5e6-483a-a508-ba36360a5ff1" (UID: "82c98e65-c5e6-483a-a508-ba36360a5ff1"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.061961 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/82c98e65-c5e6-483a-a508-ba36360a5ff1-httpd-run\") pod \"82c98e65-c5e6-483a-a508-ba36360a5ff1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.062156 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82c98e65-c5e6-483a-a508-ba36360a5ff1-config-data\") pod \"82c98e65-c5e6-483a-a508-ba36360a5ff1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.062224 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"82c98e65-c5e6-483a-a508-ba36360a5ff1\" (UID: \"82c98e65-c5e6-483a-a508-ba36360a5ff1\") " Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.062355 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82c98e65-c5e6-483a-a508-ba36360a5ff1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "82c98e65-c5e6-483a-a508-ba36360a5ff1" (UID: "82c98e65-c5e6-483a-a508-ba36360a5ff1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.062358 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-sys" (OuterVolumeSpecName: "sys") pod "82c98e65-c5e6-483a-a508-ba36360a5ff1" (UID: "82c98e65-c5e6-483a-a508-ba36360a5ff1"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.062545 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82c98e65-c5e6-483a-a508-ba36360a5ff1-logs" (OuterVolumeSpecName: "logs") pod "82c98e65-c5e6-483a-a508-ba36360a5ff1" (UID: "82c98e65-c5e6-483a-a508-ba36360a5ff1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.063247 4802 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-sys\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.063301 4802 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.063328 4802 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-dev\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.063351 4802 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.063373 4802 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/82c98e65-c5e6-483a-a508-ba36360a5ff1-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.063396 4802 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.063418 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82c98e65-c5e6-483a-a508-ba36360a5ff1-logs\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.063440 4802 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.061852 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "82c98e65-c5e6-483a-a508-ba36360a5ff1" (UID: "82c98e65-c5e6-483a-a508-ba36360a5ff1"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.068205 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "82c98e65-c5e6-483a-a508-ba36360a5ff1" (UID: "82c98e65-c5e6-483a-a508-ba36360a5ff1"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.068240 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82c98e65-c5e6-483a-a508-ba36360a5ff1-kube-api-access-j2kv6" (OuterVolumeSpecName: "kube-api-access-j2kv6") pod "82c98e65-c5e6-483a-a508-ba36360a5ff1" (UID: "82c98e65-c5e6-483a-a508-ba36360a5ff1"). InnerVolumeSpecName "kube-api-access-j2kv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.068238 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c98e65-c5e6-483a-a508-ba36360a5ff1-scripts" (OuterVolumeSpecName: "scripts") pod "82c98e65-c5e6-483a-a508-ba36360a5ff1" (UID: "82c98e65-c5e6-483a-a508-ba36360a5ff1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.068336 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "82c98e65-c5e6-483a-a508-ba36360a5ff1" (UID: "82c98e65-c5e6-483a-a508-ba36360a5ff1"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.138407 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c98e65-c5e6-483a-a508-ba36360a5ff1-config-data" (OuterVolumeSpecName: "config-data") pod "82c98e65-c5e6-483a-a508-ba36360a5ff1" (UID: "82c98e65-c5e6-483a-a508-ba36360a5ff1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.165400 4802 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/82c98e65-c5e6-483a-a508-ba36360a5ff1-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.165456 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2kv6\" (UniqueName: \"kubernetes.io/projected/82c98e65-c5e6-483a-a508-ba36360a5ff1-kube-api-access-j2kv6\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.165502 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.165519 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82c98e65-c5e6-483a-a508-ba36360a5ff1-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.165537 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.165550 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82c98e65-c5e6-483a-a508-ba36360a5ff1-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.179981 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.182382 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.266892 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.266925 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.664402 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-wc8x4"] Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.672178 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-wc8x4"] Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.727272 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glanced3b4-account-delete-rvfm5"] Nov 25 17:18:05 crc kubenswrapper[4802]: E1125 17:18:05.727612 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16385bba-8b5a-42be-9357-bb2439bf2dc4" containerName="registry-server" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.727638 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="16385bba-8b5a-42be-9357-bb2439bf2dc4" containerName="registry-server" Nov 25 17:18:05 crc kubenswrapper[4802]: E1125 17:18:05.727651 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16385bba-8b5a-42be-9357-bb2439bf2dc4" containerName="extract-utilities" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.727660 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="16385bba-8b5a-42be-9357-bb2439bf2dc4" containerName="extract-utilities" Nov 25 17:18:05 crc kubenswrapper[4802]: E1125 17:18:05.727675 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c98e65-c5e6-483a-a508-ba36360a5ff1" containerName="glance-api" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.727683 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c98e65-c5e6-483a-a508-ba36360a5ff1" containerName="glance-api" Nov 25 17:18:05 crc kubenswrapper[4802]: E1125 17:18:05.727700 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c98e65-c5e6-483a-a508-ba36360a5ff1" containerName="glance-log" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.727708 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c98e65-c5e6-483a-a508-ba36360a5ff1" containerName="glance-log" Nov 25 17:18:05 crc kubenswrapper[4802]: E1125 17:18:05.727728 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c98e65-c5e6-483a-a508-ba36360a5ff1" containerName="glance-httpd" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.727735 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c98e65-c5e6-483a-a508-ba36360a5ff1" containerName="glance-httpd" Nov 25 17:18:05 crc kubenswrapper[4802]: E1125 17:18:05.727756 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16385bba-8b5a-42be-9357-bb2439bf2dc4" containerName="extract-content" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.727764 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="16385bba-8b5a-42be-9357-bb2439bf2dc4" containerName="extract-content" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.727920 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="16385bba-8b5a-42be-9357-bb2439bf2dc4" containerName="registry-server" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.727939 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="82c98e65-c5e6-483a-a508-ba36360a5ff1" containerName="glance-httpd" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.727953 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="82c98e65-c5e6-483a-a508-ba36360a5ff1" containerName="glance-api" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.727977 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="82c98e65-c5e6-483a-a508-ba36360a5ff1" containerName="glance-log" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.728666 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced3b4-account-delete-rvfm5" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.735099 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glanced3b4-account-delete-rvfm5"] Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.794432 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"82c98e65-c5e6-483a-a508-ba36360a5ff1","Type":"ContainerDied","Data":"f25aaa7da5481cc8e6454e9af04b8c230d99038a3aa4fbff10d2c628f4adc74e"} Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.794513 4802 scope.go:117] "RemoveContainer" containerID="68aac35737ade4b501624db70503535aff2e798666206c03c83715a2b1ab2bd4" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.794521 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.825594 4802 scope.go:117] "RemoveContainer" containerID="9abaa4cffc0151fd5b5b5a0449390ac0ac0795fb6a9bb4fbc535f3584848a5f2" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.825760 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.826109 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="8ccfbe24-ca2d-447d-9675-6836cd5b02b3" containerName="glance-log" containerID="cri-o://e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b" gracePeriod=30 Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.826334 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="8ccfbe24-ca2d-447d-9675-6836cd5b02b3" containerName="glance-httpd" containerID="cri-o://8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e" gracePeriod=30 Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.826308 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="8ccfbe24-ca2d-447d-9675-6836cd5b02b3" containerName="glance-api" containerID="cri-o://b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051" gracePeriod=30 Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.844075 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.851424 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.856223 4802 scope.go:117] "RemoveContainer" containerID="edc8ae7d5b99e2e0e0496e99d6c2051e85d188fbdaf35a5bc0ecac5849c93f08" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.875213 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47c59d93-2330-4d02-9f09-0c812d220e3f-operator-scripts\") pod \"glanced3b4-account-delete-rvfm5\" (UID: \"47c59d93-2330-4d02-9f09-0c812d220e3f\") " pod="glance-kuttl-tests/glanced3b4-account-delete-rvfm5" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.875248 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c7kt\" (UniqueName: \"kubernetes.io/projected/47c59d93-2330-4d02-9f09-0c812d220e3f-kube-api-access-2c7kt\") pod \"glanced3b4-account-delete-rvfm5\" (UID: \"47c59d93-2330-4d02-9f09-0c812d220e3f\") " pod="glance-kuttl-tests/glanced3b4-account-delete-rvfm5" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.905881 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.906178 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="1f9ba153-fd78-4203-b36f-39e14d80e747" containerName="glance-log" containerID="cri-o://fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef" gracePeriod=30 Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.906526 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="1f9ba153-fd78-4203-b36f-39e14d80e747" containerName="glance-api" containerID="cri-o://27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6" gracePeriod=30 Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.906575 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="1f9ba153-fd78-4203-b36f-39e14d80e747" containerName="glance-httpd" containerID="cri-o://caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca" gracePeriod=30 Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.976714 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47c59d93-2330-4d02-9f09-0c812d220e3f-operator-scripts\") pod \"glanced3b4-account-delete-rvfm5\" (UID: \"47c59d93-2330-4d02-9f09-0c812d220e3f\") " pod="glance-kuttl-tests/glanced3b4-account-delete-rvfm5" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.976753 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c7kt\" (UniqueName: \"kubernetes.io/projected/47c59d93-2330-4d02-9f09-0c812d220e3f-kube-api-access-2c7kt\") pod \"glanced3b4-account-delete-rvfm5\" (UID: \"47c59d93-2330-4d02-9f09-0c812d220e3f\") " pod="glance-kuttl-tests/glanced3b4-account-delete-rvfm5" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.977868 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47c59d93-2330-4d02-9f09-0c812d220e3f-operator-scripts\") pod \"glanced3b4-account-delete-rvfm5\" (UID: \"47c59d93-2330-4d02-9f09-0c812d220e3f\") " pod="glance-kuttl-tests/glanced3b4-account-delete-rvfm5" Nov 25 17:18:05 crc kubenswrapper[4802]: I1125 17:18:05.994855 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c7kt\" (UniqueName: \"kubernetes.io/projected/47c59d93-2330-4d02-9f09-0c812d220e3f-kube-api-access-2c7kt\") pod \"glanced3b4-account-delete-rvfm5\" (UID: \"47c59d93-2330-4d02-9f09-0c812d220e3f\") " pod="glance-kuttl-tests/glanced3b4-account-delete-rvfm5" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.043213 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced3b4-account-delete-rvfm5" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.523844 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glanced3b4-account-delete-rvfm5"] Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.650892 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.687934 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.786797 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.786872 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-sys\") pod \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.786904 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-lib-modules\") pod \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.786947 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-logs\") pod \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.786970 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-config-data\") pod \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787027 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-sys" (OuterVolumeSpecName: "sys") pod "8ccfbe24-ca2d-447d-9675-6836cd5b02b3" (UID: "8ccfbe24-ca2d-447d-9675-6836cd5b02b3"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787041 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "8ccfbe24-ca2d-447d-9675-6836cd5b02b3" (UID: "8ccfbe24-ca2d-447d-9675-6836cd5b02b3"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787050 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787112 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6wtz\" (UniqueName: \"kubernetes.io/projected/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-kube-api-access-q6wtz\") pod \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787168 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-run\") pod \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787194 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-scripts\") pod \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787236 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-var-locks-brick\") pod \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787259 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-httpd-run\") pod \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787287 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-etc-iscsi\") pod \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787315 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-etc-nvme\") pod \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787341 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-dev\") pod \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\" (UID: \"8ccfbe24-ca2d-447d-9675-6836cd5b02b3\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787300 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-run" (OuterVolumeSpecName: "run") pod "8ccfbe24-ca2d-447d-9675-6836cd5b02b3" (UID: "8ccfbe24-ca2d-447d-9675-6836cd5b02b3"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787335 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "8ccfbe24-ca2d-447d-9675-6836cd5b02b3" (UID: "8ccfbe24-ca2d-447d-9675-6836cd5b02b3"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787758 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "8ccfbe24-ca2d-447d-9675-6836cd5b02b3" (UID: "8ccfbe24-ca2d-447d-9675-6836cd5b02b3"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787545 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-dev" (OuterVolumeSpecName: "dev") pod "8ccfbe24-ca2d-447d-9675-6836cd5b02b3" (UID: "8ccfbe24-ca2d-447d-9675-6836cd5b02b3"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787722 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "8ccfbe24-ca2d-447d-9675-6836cd5b02b3" (UID: "8ccfbe24-ca2d-447d-9675-6836cd5b02b3"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787733 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8ccfbe24-ca2d-447d-9675-6836cd5b02b3" (UID: "8ccfbe24-ca2d-447d-9675-6836cd5b02b3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787895 4802 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787913 4802 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787923 4802 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787932 4802 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787941 4802 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787950 4802 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-dev\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787959 4802 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-sys\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787967 4802 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.787994 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-logs" (OuterVolumeSpecName: "logs") pod "8ccfbe24-ca2d-447d-9675-6836cd5b02b3" (UID: "8ccfbe24-ca2d-447d-9675-6836cd5b02b3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.795823 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "8ccfbe24-ca2d-447d-9675-6836cd5b02b3" (UID: "8ccfbe24-ca2d-447d-9675-6836cd5b02b3"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.797186 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance-cache") pod "8ccfbe24-ca2d-447d-9675-6836cd5b02b3" (UID: "8ccfbe24-ca2d-447d-9675-6836cd5b02b3"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.797947 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-scripts" (OuterVolumeSpecName: "scripts") pod "8ccfbe24-ca2d-447d-9675-6836cd5b02b3" (UID: "8ccfbe24-ca2d-447d-9675-6836cd5b02b3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.797953 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-kube-api-access-q6wtz" (OuterVolumeSpecName: "kube-api-access-q6wtz") pod "8ccfbe24-ca2d-447d-9675-6836cd5b02b3" (UID: "8ccfbe24-ca2d-447d-9675-6836cd5b02b3"). InnerVolumeSpecName "kube-api-access-q6wtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.808166 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1f9ba153-fd78-4203-b36f-39e14d80e747","Type":"ContainerDied","Data":"27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6"} Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.808218 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.808238 4802 scope.go:117] "RemoveContainer" containerID="27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.808101 4802 generic.go:334] "Generic (PLEG): container finished" podID="1f9ba153-fd78-4203-b36f-39e14d80e747" containerID="27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6" exitCode=0 Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.808281 4802 generic.go:334] "Generic (PLEG): container finished" podID="1f9ba153-fd78-4203-b36f-39e14d80e747" containerID="caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca" exitCode=0 Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.808289 4802 generic.go:334] "Generic (PLEG): container finished" podID="1f9ba153-fd78-4203-b36f-39e14d80e747" containerID="fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef" exitCode=143 Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.808382 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1f9ba153-fd78-4203-b36f-39e14d80e747","Type":"ContainerDied","Data":"caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca"} Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.808444 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1f9ba153-fd78-4203-b36f-39e14d80e747","Type":"ContainerDied","Data":"fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef"} Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.808455 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1f9ba153-fd78-4203-b36f-39e14d80e747","Type":"ContainerDied","Data":"6c25a06e246e174a57dd57e540afd10843aa870584b2aa2a5d4dbec3b20dbb55"} Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.810602 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanced3b4-account-delete-rvfm5" event={"ID":"47c59d93-2330-4d02-9f09-0c812d220e3f","Type":"ContainerStarted","Data":"1a4940f0df0c4c4f6bb47145dfcfd12ee7b5dee23fd5ab9b7acdd0bd510c2ce8"} Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.810637 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanced3b4-account-delete-rvfm5" event={"ID":"47c59d93-2330-4d02-9f09-0c812d220e3f","Type":"ContainerStarted","Data":"708f7872b9e1b078e4c27a23f7a2c96772b4e4aeec08f8d7ef1a46961b54a203"} Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.816323 4802 generic.go:334] "Generic (PLEG): container finished" podID="8ccfbe24-ca2d-447d-9675-6836cd5b02b3" containerID="b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051" exitCode=0 Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.816363 4802 generic.go:334] "Generic (PLEG): container finished" podID="8ccfbe24-ca2d-447d-9675-6836cd5b02b3" containerID="8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e" exitCode=0 Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.816375 4802 generic.go:334] "Generic (PLEG): container finished" podID="8ccfbe24-ca2d-447d-9675-6836cd5b02b3" containerID="e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b" exitCode=143 Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.816435 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8ccfbe24-ca2d-447d-9675-6836cd5b02b3","Type":"ContainerDied","Data":"b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051"} Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.816468 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8ccfbe24-ca2d-447d-9675-6836cd5b02b3","Type":"ContainerDied","Data":"8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e"} Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.816483 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8ccfbe24-ca2d-447d-9675-6836cd5b02b3","Type":"ContainerDied","Data":"e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b"} Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.816496 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8ccfbe24-ca2d-447d-9675-6836cd5b02b3","Type":"ContainerDied","Data":"9aed5f3020cd19bf2b496b3f16cb1288a9376c6bd50fcae09eac6ea1631eb731"} Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.816568 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.829081 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glanced3b4-account-delete-rvfm5" podStartSLOduration=1.8290566400000001 podStartE2EDuration="1.82905664s" podCreationTimestamp="2025-11-25 17:18:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 17:18:06.824595188 +0000 UTC m=+1869.968942374" watchObservedRunningTime="2025-11-25 17:18:06.82905664 +0000 UTC m=+1869.973403846" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.836464 4802 scope.go:117] "RemoveContainer" containerID="caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.856265 4802 scope.go:117] "RemoveContainer" containerID="fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.858836 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-config-data" (OuterVolumeSpecName: "config-data") pod "8ccfbe24-ca2d-447d-9675-6836cd5b02b3" (UID: "8ccfbe24-ca2d-447d-9675-6836cd5b02b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.878922 4802 scope.go:117] "RemoveContainer" containerID="27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6" Nov 25 17:18:06 crc kubenswrapper[4802]: E1125 17:18:06.881478 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6\": container with ID starting with 27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6 not found: ID does not exist" containerID="27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.881515 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6"} err="failed to get container status \"27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6\": rpc error: code = NotFound desc = could not find container \"27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6\": container with ID starting with 27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6 not found: ID does not exist" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.881537 4802 scope.go:117] "RemoveContainer" containerID="caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca" Nov 25 17:18:06 crc kubenswrapper[4802]: E1125 17:18:06.881996 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca\": container with ID starting with caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca not found: ID does not exist" containerID="caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.882022 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca"} err="failed to get container status \"caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca\": rpc error: code = NotFound desc = could not find container \"caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca\": container with ID starting with caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca not found: ID does not exist" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.882036 4802 scope.go:117] "RemoveContainer" containerID="fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef" Nov 25 17:18:06 crc kubenswrapper[4802]: E1125 17:18:06.883092 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef\": container with ID starting with fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef not found: ID does not exist" containerID="fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.883158 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef"} err="failed to get container status \"fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef\": rpc error: code = NotFound desc = could not find container \"fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef\": container with ID starting with fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef not found: ID does not exist" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.883193 4802 scope.go:117] "RemoveContainer" containerID="27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.883534 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6"} err="failed to get container status \"27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6\": rpc error: code = NotFound desc = could not find container \"27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6\": container with ID starting with 27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6 not found: ID does not exist" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.883558 4802 scope.go:117] "RemoveContainer" containerID="caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.883875 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca"} err="failed to get container status \"caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca\": rpc error: code = NotFound desc = could not find container \"caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca\": container with ID starting with caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca not found: ID does not exist" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.883904 4802 scope.go:117] "RemoveContainer" containerID="fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.884376 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef"} err="failed to get container status \"fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef\": rpc error: code = NotFound desc = could not find container \"fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef\": container with ID starting with fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef not found: ID does not exist" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.884419 4802 scope.go:117] "RemoveContainer" containerID="27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.884721 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6"} err="failed to get container status \"27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6\": rpc error: code = NotFound desc = could not find container \"27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6\": container with ID starting with 27503e3d62b095dcb8dc029ec0c6d009ddfc95c218b39dc4db0aac685d6ad8a6 not found: ID does not exist" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.884751 4802 scope.go:117] "RemoveContainer" containerID="caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.884983 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca"} err="failed to get container status \"caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca\": rpc error: code = NotFound desc = could not find container \"caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca\": container with ID starting with caa2d0dfc27625389e52ef83c07bfe2b7513810064eb99b0e99419b29e3917ca not found: ID does not exist" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.885005 4802 scope.go:117] "RemoveContainer" containerID="fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.885275 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef"} err="failed to get container status \"fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef\": rpc error: code = NotFound desc = could not find container \"fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef\": container with ID starting with fb3d0126a1b46aa30e334a38148ee7cebb188ee28048506183a1029fce7ae2ef not found: ID does not exist" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.885301 4802 scope.go:117] "RemoveContainer" containerID="b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.889455 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f9ba153-fd78-4203-b36f-39e14d80e747-httpd-run\") pod \"1f9ba153-fd78-4203-b36f-39e14d80e747\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.889496 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f9ba153-fd78-4203-b36f-39e14d80e747-config-data\") pod \"1f9ba153-fd78-4203-b36f-39e14d80e747\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.889521 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-sys\") pod \"1f9ba153-fd78-4203-b36f-39e14d80e747\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.889557 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjvzp\" (UniqueName: \"kubernetes.io/projected/1f9ba153-fd78-4203-b36f-39e14d80e747-kube-api-access-mjvzp\") pod \"1f9ba153-fd78-4203-b36f-39e14d80e747\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.889626 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-run\") pod \"1f9ba153-fd78-4203-b36f-39e14d80e747\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.889669 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-etc-iscsi\") pod \"1f9ba153-fd78-4203-b36f-39e14d80e747\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.889693 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-etc-nvme\") pod \"1f9ba153-fd78-4203-b36f-39e14d80e747\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.889713 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-lib-modules\") pod \"1f9ba153-fd78-4203-b36f-39e14d80e747\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.889769 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f9ba153-fd78-4203-b36f-39e14d80e747-logs\") pod \"1f9ba153-fd78-4203-b36f-39e14d80e747\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.889802 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f9ba153-fd78-4203-b36f-39e14d80e747-scripts\") pod \"1f9ba153-fd78-4203-b36f-39e14d80e747\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.889828 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-dev\") pod \"1f9ba153-fd78-4203-b36f-39e14d80e747\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.889847 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-var-locks-brick\") pod \"1f9ba153-fd78-4203-b36f-39e14d80e747\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.889900 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"1f9ba153-fd78-4203-b36f-39e14d80e747\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.889898 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "1f9ba153-fd78-4203-b36f-39e14d80e747" (UID: "1f9ba153-fd78-4203-b36f-39e14d80e747"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.889920 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"1f9ba153-fd78-4203-b36f-39e14d80e747\" (UID: \"1f9ba153-fd78-4203-b36f-39e14d80e747\") " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.889945 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-run" (OuterVolumeSpecName: "run") pod "1f9ba153-fd78-4203-b36f-39e14d80e747" (UID: "1f9ba153-fd78-4203-b36f-39e14d80e747"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.889971 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "1f9ba153-fd78-4203-b36f-39e14d80e747" (UID: "1f9ba153-fd78-4203-b36f-39e14d80e747"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.890211 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-sys" (OuterVolumeSpecName: "sys") pod "1f9ba153-fd78-4203-b36f-39e14d80e747" (UID: "1f9ba153-fd78-4203-b36f-39e14d80e747"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.890578 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.890611 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-logs\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.890606 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f9ba153-fd78-4203-b36f-39e14d80e747-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1f9ba153-fd78-4203-b36f-39e14d80e747" (UID: "1f9ba153-fd78-4203-b36f-39e14d80e747"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.890626 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.890735 4802 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-sys\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.890780 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.890803 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6wtz\" (UniqueName: \"kubernetes.io/projected/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-kube-api-access-q6wtz\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.890827 4802 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.890845 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ccfbe24-ca2d-447d-9675-6836cd5b02b3-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.890862 4802 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.890881 4802 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.891161 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "1f9ba153-fd78-4203-b36f-39e14d80e747" (UID: "1f9ba153-fd78-4203-b36f-39e14d80e747"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.891199 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "1f9ba153-fd78-4203-b36f-39e14d80e747" (UID: "1f9ba153-fd78-4203-b36f-39e14d80e747"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.891252 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-dev" (OuterVolumeSpecName: "dev") pod "1f9ba153-fd78-4203-b36f-39e14d80e747" (UID: "1f9ba153-fd78-4203-b36f-39e14d80e747"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.891694 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f9ba153-fd78-4203-b36f-39e14d80e747-logs" (OuterVolumeSpecName: "logs") pod "1f9ba153-fd78-4203-b36f-39e14d80e747" (UID: "1f9ba153-fd78-4203-b36f-39e14d80e747"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.893956 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f9ba153-fd78-4203-b36f-39e14d80e747-kube-api-access-mjvzp" (OuterVolumeSpecName: "kube-api-access-mjvzp") pod "1f9ba153-fd78-4203-b36f-39e14d80e747" (UID: "1f9ba153-fd78-4203-b36f-39e14d80e747"). InnerVolumeSpecName "kube-api-access-mjvzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.895290 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance-cache") pod "1f9ba153-fd78-4203-b36f-39e14d80e747" (UID: "1f9ba153-fd78-4203-b36f-39e14d80e747"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.896462 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "1f9ba153-fd78-4203-b36f-39e14d80e747" (UID: "1f9ba153-fd78-4203-b36f-39e14d80e747"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.896564 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f9ba153-fd78-4203-b36f-39e14d80e747-scripts" (OuterVolumeSpecName: "scripts") pod "1f9ba153-fd78-4203-b36f-39e14d80e747" (UID: "1f9ba153-fd78-4203-b36f-39e14d80e747"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.911103 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.918834 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.957139 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f9ba153-fd78-4203-b36f-39e14d80e747-config-data" (OuterVolumeSpecName: "config-data") pod "1f9ba153-fd78-4203-b36f-39e14d80e747" (UID: "1f9ba153-fd78-4203-b36f-39e14d80e747"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.960049 4802 scope.go:117] "RemoveContainer" containerID="8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.979134 4802 scope.go:117] "RemoveContainer" containerID="e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.991344 4802 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f9ba153-fd78-4203-b36f-39e14d80e747-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.991373 4802 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f9ba153-fd78-4203-b36f-39e14d80e747-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.991389 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjvzp\" (UniqueName: \"kubernetes.io/projected/1f9ba153-fd78-4203-b36f-39e14d80e747-kube-api-access-mjvzp\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.991405 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.991419 4802 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.991430 4802 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f9ba153-fd78-4203-b36f-39e14d80e747-logs\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.991440 4802 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f9ba153-fd78-4203-b36f-39e14d80e747-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.991450 4802 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-dev\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.991461 4802 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1f9ba153-fd78-4203-b36f-39e14d80e747-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.991501 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.991531 4802 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.991546 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.995801 4802 scope.go:117] "RemoveContainer" containerID="b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051" Nov 25 17:18:06 crc kubenswrapper[4802]: E1125 17:18:06.996312 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051\": container with ID starting with b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051 not found: ID does not exist" containerID="b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.996344 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051"} err="failed to get container status \"b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051\": rpc error: code = NotFound desc = could not find container \"b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051\": container with ID starting with b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051 not found: ID does not exist" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.996370 4802 scope.go:117] "RemoveContainer" containerID="8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e" Nov 25 17:18:06 crc kubenswrapper[4802]: E1125 17:18:06.996590 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e\": container with ID starting with 8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e not found: ID does not exist" containerID="8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.996609 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e"} err="failed to get container status \"8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e\": rpc error: code = NotFound desc = could not find container \"8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e\": container with ID starting with 8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e not found: ID does not exist" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.996620 4802 scope.go:117] "RemoveContainer" containerID="e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b" Nov 25 17:18:06 crc kubenswrapper[4802]: E1125 17:18:06.996799 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b\": container with ID starting with e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b not found: ID does not exist" containerID="e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.996817 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b"} err="failed to get container status \"e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b\": rpc error: code = NotFound desc = could not find container \"e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b\": container with ID starting with e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b not found: ID does not exist" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.996828 4802 scope.go:117] "RemoveContainer" containerID="b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.997764 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051"} err="failed to get container status \"b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051\": rpc error: code = NotFound desc = could not find container \"b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051\": container with ID starting with b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051 not found: ID does not exist" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.997783 4802 scope.go:117] "RemoveContainer" containerID="8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.997989 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e"} err="failed to get container status \"8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e\": rpc error: code = NotFound desc = could not find container \"8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e\": container with ID starting with 8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e not found: ID does not exist" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.998031 4802 scope.go:117] "RemoveContainer" containerID="e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.998286 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b"} err="failed to get container status \"e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b\": rpc error: code = NotFound desc = could not find container \"e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b\": container with ID starting with e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b not found: ID does not exist" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.998304 4802 scope.go:117] "RemoveContainer" containerID="b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.998535 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051"} err="failed to get container status \"b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051\": rpc error: code = NotFound desc = could not find container \"b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051\": container with ID starting with b1ded792f22c68e3a27343351e8c6292259a35f9d18afc6fac61392b07fed051 not found: ID does not exist" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.998554 4802 scope.go:117] "RemoveContainer" containerID="8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.998783 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e"} err="failed to get container status \"8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e\": rpc error: code = NotFound desc = could not find container \"8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e\": container with ID starting with 8c4ff24a15b675b5561fd41b5df1674212e86fdd943293bbe2d2b5b7b6aa9e6e not found: ID does not exist" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.998799 4802 scope.go:117] "RemoveContainer" containerID="e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b" Nov 25 17:18:06 crc kubenswrapper[4802]: I1125 17:18:06.998997 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b"} err="failed to get container status \"e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b\": rpc error: code = NotFound desc = could not find container \"e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b\": container with ID starting with e8b598c9425439b003e43bab7a3cf7b7c6352b6a5f9805ad0d7248095994ce3b not found: ID does not exist" Nov 25 17:18:07 crc kubenswrapper[4802]: I1125 17:18:07.003680 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 25 17:18:07 crc kubenswrapper[4802]: I1125 17:18:07.004272 4802 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 25 17:18:07 crc kubenswrapper[4802]: I1125 17:18:07.092514 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:07 crc kubenswrapper[4802]: I1125 17:18:07.092543 4802 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:07 crc kubenswrapper[4802]: I1125 17:18:07.147979 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 17:18:07 crc kubenswrapper[4802]: I1125 17:18:07.165972 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 17:18:07 crc kubenswrapper[4802]: I1125 17:18:07.185900 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 17:18:07 crc kubenswrapper[4802]: I1125 17:18:07.192911 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 17:18:07 crc kubenswrapper[4802]: I1125 17:18:07.511658 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f9ba153-fd78-4203-b36f-39e14d80e747" path="/var/lib/kubelet/pods/1f9ba153-fd78-4203-b36f-39e14d80e747/volumes" Nov 25 17:18:07 crc kubenswrapper[4802]: I1125 17:18:07.512298 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dde71ad-bcda-407e-a490-e8323685e7eb" path="/var/lib/kubelet/pods/2dde71ad-bcda-407e-a490-e8323685e7eb/volumes" Nov 25 17:18:07 crc kubenswrapper[4802]: I1125 17:18:07.512952 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82c98e65-c5e6-483a-a508-ba36360a5ff1" path="/var/lib/kubelet/pods/82c98e65-c5e6-483a-a508-ba36360a5ff1/volumes" Nov 25 17:18:07 crc kubenswrapper[4802]: I1125 17:18:07.513594 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ccfbe24-ca2d-447d-9675-6836cd5b02b3" path="/var/lib/kubelet/pods/8ccfbe24-ca2d-447d-9675-6836cd5b02b3/volumes" Nov 25 17:18:07 crc kubenswrapper[4802]: I1125 17:18:07.833630 4802 generic.go:334] "Generic (PLEG): container finished" podID="47c59d93-2330-4d02-9f09-0c812d220e3f" containerID="1a4940f0df0c4c4f6bb47145dfcfd12ee7b5dee23fd5ab9b7acdd0bd510c2ce8" exitCode=0 Nov 25 17:18:07 crc kubenswrapper[4802]: I1125 17:18:07.833734 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanced3b4-account-delete-rvfm5" event={"ID":"47c59d93-2330-4d02-9f09-0c812d220e3f","Type":"ContainerDied","Data":"1a4940f0df0c4c4f6bb47145dfcfd12ee7b5dee23fd5ab9b7acdd0bd510c2ce8"} Nov 25 17:18:09 crc kubenswrapper[4802]: I1125 17:18:09.154736 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced3b4-account-delete-rvfm5" Nov 25 17:18:09 crc kubenswrapper[4802]: I1125 17:18:09.323466 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2c7kt\" (UniqueName: \"kubernetes.io/projected/47c59d93-2330-4d02-9f09-0c812d220e3f-kube-api-access-2c7kt\") pod \"47c59d93-2330-4d02-9f09-0c812d220e3f\" (UID: \"47c59d93-2330-4d02-9f09-0c812d220e3f\") " Nov 25 17:18:09 crc kubenswrapper[4802]: I1125 17:18:09.323529 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47c59d93-2330-4d02-9f09-0c812d220e3f-operator-scripts\") pod \"47c59d93-2330-4d02-9f09-0c812d220e3f\" (UID: \"47c59d93-2330-4d02-9f09-0c812d220e3f\") " Nov 25 17:18:09 crc kubenswrapper[4802]: I1125 17:18:09.324540 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47c59d93-2330-4d02-9f09-0c812d220e3f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "47c59d93-2330-4d02-9f09-0c812d220e3f" (UID: "47c59d93-2330-4d02-9f09-0c812d220e3f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 17:18:09 crc kubenswrapper[4802]: I1125 17:18:09.329302 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47c59d93-2330-4d02-9f09-0c812d220e3f-kube-api-access-2c7kt" (OuterVolumeSpecName: "kube-api-access-2c7kt") pod "47c59d93-2330-4d02-9f09-0c812d220e3f" (UID: "47c59d93-2330-4d02-9f09-0c812d220e3f"). InnerVolumeSpecName "kube-api-access-2c7kt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:18:09 crc kubenswrapper[4802]: I1125 17:18:09.425044 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2c7kt\" (UniqueName: \"kubernetes.io/projected/47c59d93-2330-4d02-9f09-0c812d220e3f-kube-api-access-2c7kt\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:09 crc kubenswrapper[4802]: I1125 17:18:09.425080 4802 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47c59d93-2330-4d02-9f09-0c812d220e3f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 17:18:09 crc kubenswrapper[4802]: I1125 17:18:09.853620 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanced3b4-account-delete-rvfm5" event={"ID":"47c59d93-2330-4d02-9f09-0c812d220e3f","Type":"ContainerDied","Data":"708f7872b9e1b078e4c27a23f7a2c96772b4e4aeec08f8d7ef1a46961b54a203"} Nov 25 17:18:09 crc kubenswrapper[4802]: I1125 17:18:09.853674 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="708f7872b9e1b078e4c27a23f7a2c96772b4e4aeec08f8d7ef1a46961b54a203" Nov 25 17:18:09 crc kubenswrapper[4802]: I1125 17:18:09.853689 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced3b4-account-delete-rvfm5" Nov 25 17:18:10 crc kubenswrapper[4802]: I1125 17:18:10.734544 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-qbqgx"] Nov 25 17:18:10 crc kubenswrapper[4802]: I1125 17:18:10.745193 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-qbqgx"] Nov 25 17:18:10 crc kubenswrapper[4802]: I1125 17:18:10.752814 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr"] Nov 25 17:18:10 crc kubenswrapper[4802]: I1125 17:18:10.758468 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glanced3b4-account-delete-rvfm5"] Nov 25 17:18:10 crc kubenswrapper[4802]: I1125 17:18:10.763843 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-d3b4-account-create-update-h4wcr"] Nov 25 17:18:10 crc kubenswrapper[4802]: I1125 17:18:10.769074 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glanced3b4-account-delete-rvfm5"] Nov 25 17:18:11 crc kubenswrapper[4802]: I1125 17:18:11.520044 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47c59d93-2330-4d02-9f09-0c812d220e3f" path="/var/lib/kubelet/pods/47c59d93-2330-4d02-9f09-0c812d220e3f/volumes" Nov 25 17:18:11 crc kubenswrapper[4802]: I1125 17:18:11.521092 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a38f8ea1-50bb-4b06-883c-9be588679c41" path="/var/lib/kubelet/pods/a38f8ea1-50bb-4b06-883c-9be588679c41/volumes" Nov 25 17:18:11 crc kubenswrapper[4802]: I1125 17:18:11.522115 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e34282e2-a266-4d67-ae13-0bce1083ae91" path="/var/lib/kubelet/pods/e34282e2-a266-4d67-ae13-0bce1083ae91/volumes" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.820954 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-nn27p/must-gather-spl62"] Nov 25 17:18:35 crc kubenswrapper[4802]: E1125 17:18:35.821893 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47c59d93-2330-4d02-9f09-0c812d220e3f" containerName="mariadb-account-delete" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.821912 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="47c59d93-2330-4d02-9f09-0c812d220e3f" containerName="mariadb-account-delete" Nov 25 17:18:35 crc kubenswrapper[4802]: E1125 17:18:35.821932 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f9ba153-fd78-4203-b36f-39e14d80e747" containerName="glance-api" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.821940 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f9ba153-fd78-4203-b36f-39e14d80e747" containerName="glance-api" Nov 25 17:18:35 crc kubenswrapper[4802]: E1125 17:18:35.821967 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f9ba153-fd78-4203-b36f-39e14d80e747" containerName="glance-log" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.821975 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f9ba153-fd78-4203-b36f-39e14d80e747" containerName="glance-log" Nov 25 17:18:35 crc kubenswrapper[4802]: E1125 17:18:35.821991 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ccfbe24-ca2d-447d-9675-6836cd5b02b3" containerName="glance-log" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.821998 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ccfbe24-ca2d-447d-9675-6836cd5b02b3" containerName="glance-log" Nov 25 17:18:35 crc kubenswrapper[4802]: E1125 17:18:35.822010 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ccfbe24-ca2d-447d-9675-6836cd5b02b3" containerName="glance-api" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.822019 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ccfbe24-ca2d-447d-9675-6836cd5b02b3" containerName="glance-api" Nov 25 17:18:35 crc kubenswrapper[4802]: E1125 17:18:35.822034 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f9ba153-fd78-4203-b36f-39e14d80e747" containerName="glance-httpd" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.822042 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f9ba153-fd78-4203-b36f-39e14d80e747" containerName="glance-httpd" Nov 25 17:18:35 crc kubenswrapper[4802]: E1125 17:18:35.822057 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ccfbe24-ca2d-447d-9675-6836cd5b02b3" containerName="glance-httpd" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.822064 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ccfbe24-ca2d-447d-9675-6836cd5b02b3" containerName="glance-httpd" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.822234 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ccfbe24-ca2d-447d-9675-6836cd5b02b3" containerName="glance-log" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.822256 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="47c59d93-2330-4d02-9f09-0c812d220e3f" containerName="mariadb-account-delete" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.822271 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ccfbe24-ca2d-447d-9675-6836cd5b02b3" containerName="glance-httpd" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.822283 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f9ba153-fd78-4203-b36f-39e14d80e747" containerName="glance-log" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.822295 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f9ba153-fd78-4203-b36f-39e14d80e747" containerName="glance-api" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.822309 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ccfbe24-ca2d-447d-9675-6836cd5b02b3" containerName="glance-api" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.822319 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f9ba153-fd78-4203-b36f-39e14d80e747" containerName="glance-httpd" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.823134 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nn27p/must-gather-spl62" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.826291 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-nn27p"/"kube-root-ca.crt" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.826508 4802 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-nn27p"/"openshift-service-ca.crt" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.837792 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-nn27p/must-gather-spl62"] Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.849325 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmrjb\" (UniqueName: \"kubernetes.io/projected/a2c3e739-8ad9-4d34-a24e-8f4b204d05f1-kube-api-access-kmrjb\") pod \"must-gather-spl62\" (UID: \"a2c3e739-8ad9-4d34-a24e-8f4b204d05f1\") " pod="openshift-must-gather-nn27p/must-gather-spl62" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.849700 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a2c3e739-8ad9-4d34-a24e-8f4b204d05f1-must-gather-output\") pod \"must-gather-spl62\" (UID: \"a2c3e739-8ad9-4d34-a24e-8f4b204d05f1\") " pod="openshift-must-gather-nn27p/must-gather-spl62" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.950763 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a2c3e739-8ad9-4d34-a24e-8f4b204d05f1-must-gather-output\") pod \"must-gather-spl62\" (UID: \"a2c3e739-8ad9-4d34-a24e-8f4b204d05f1\") " pod="openshift-must-gather-nn27p/must-gather-spl62" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.950868 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmrjb\" (UniqueName: \"kubernetes.io/projected/a2c3e739-8ad9-4d34-a24e-8f4b204d05f1-kube-api-access-kmrjb\") pod \"must-gather-spl62\" (UID: \"a2c3e739-8ad9-4d34-a24e-8f4b204d05f1\") " pod="openshift-must-gather-nn27p/must-gather-spl62" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.951246 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a2c3e739-8ad9-4d34-a24e-8f4b204d05f1-must-gather-output\") pod \"must-gather-spl62\" (UID: \"a2c3e739-8ad9-4d34-a24e-8f4b204d05f1\") " pod="openshift-must-gather-nn27p/must-gather-spl62" Nov 25 17:18:35 crc kubenswrapper[4802]: I1125 17:18:35.971216 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmrjb\" (UniqueName: \"kubernetes.io/projected/a2c3e739-8ad9-4d34-a24e-8f4b204d05f1-kube-api-access-kmrjb\") pod \"must-gather-spl62\" (UID: \"a2c3e739-8ad9-4d34-a24e-8f4b204d05f1\") " pod="openshift-must-gather-nn27p/must-gather-spl62" Nov 25 17:18:36 crc kubenswrapper[4802]: I1125 17:18:36.145289 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nn27p/must-gather-spl62" Nov 25 17:18:36 crc kubenswrapper[4802]: I1125 17:18:36.553621 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-nn27p/must-gather-spl62"] Nov 25 17:18:37 crc kubenswrapper[4802]: I1125 17:18:37.099640 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nn27p/must-gather-spl62" event={"ID":"a2c3e739-8ad9-4d34-a24e-8f4b204d05f1","Type":"ContainerStarted","Data":"ad9bbc8ca02bea2f79b57e2a06f55aded93d96e7680ce6952d41276af875dccd"} Nov 25 17:18:41 crc kubenswrapper[4802]: I1125 17:18:41.146214 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nn27p/must-gather-spl62" event={"ID":"a2c3e739-8ad9-4d34-a24e-8f4b204d05f1","Type":"ContainerStarted","Data":"c9716fe7de23debe02bc930037841704eaad1522e9e883d62c7fb56ed2ee1dbd"} Nov 25 17:18:41 crc kubenswrapper[4802]: I1125 17:18:41.146595 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nn27p/must-gather-spl62" event={"ID":"a2c3e739-8ad9-4d34-a24e-8f4b204d05f1","Type":"ContainerStarted","Data":"acea1ffae229b7bf64717bebc1f977aa771ae0ce5cf2ba32e01730de226bacd3"} Nov 25 17:18:41 crc kubenswrapper[4802]: I1125 17:18:41.170536 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-nn27p/must-gather-spl62" podStartSLOduration=2.524476762 podStartE2EDuration="6.170512688s" podCreationTimestamp="2025-11-25 17:18:35 +0000 UTC" firstStartedPulling="2025-11-25 17:18:36.566343695 +0000 UTC m=+1899.710690881" lastFinishedPulling="2025-11-25 17:18:40.212379621 +0000 UTC m=+1903.356726807" observedRunningTime="2025-11-25 17:18:41.164420782 +0000 UTC m=+1904.308767978" watchObservedRunningTime="2025-11-25 17:18:41.170512688 +0000 UTC m=+1904.314859874" Nov 25 17:19:04 crc kubenswrapper[4802]: I1125 17:19:04.497539 4802 scope.go:117] "RemoveContainer" containerID="0507e27121ac8217fc618d1b0fd8237f3553d09abcdc07eb1cb7e543c8f29df4" Nov 25 17:19:04 crc kubenswrapper[4802]: I1125 17:19:04.519395 4802 scope.go:117] "RemoveContainer" containerID="6753f6c24486761b31143a299647d5da36bb268414c2eb2817bfaa3ff1139d8e" Nov 25 17:19:04 crc kubenswrapper[4802]: I1125 17:19:04.585459 4802 scope.go:117] "RemoveContainer" containerID="ce9957fb5a14f6a8b87b508a6be9ee212b4a35a6d38b842c0f8cc619c05341e4" Nov 25 17:19:15 crc kubenswrapper[4802]: I1125 17:19:15.696942 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k_9ad4315c-9d47-485c-b656-4ba2e41803f6/util/0.log" Nov 25 17:19:15 crc kubenswrapper[4802]: I1125 17:19:15.894273 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k_9ad4315c-9d47-485c-b656-4ba2e41803f6/util/0.log" Nov 25 17:19:15 crc kubenswrapper[4802]: I1125 17:19:15.897208 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k_9ad4315c-9d47-485c-b656-4ba2e41803f6/pull/0.log" Nov 25 17:19:15 crc kubenswrapper[4802]: I1125 17:19:15.903696 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k_9ad4315c-9d47-485c-b656-4ba2e41803f6/pull/0.log" Nov 25 17:19:16 crc kubenswrapper[4802]: I1125 17:19:16.081001 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k_9ad4315c-9d47-485c-b656-4ba2e41803f6/util/0.log" Nov 25 17:19:16 crc kubenswrapper[4802]: I1125 17:19:16.139972 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k_9ad4315c-9d47-485c-b656-4ba2e41803f6/pull/0.log" Nov 25 17:19:16 crc kubenswrapper[4802]: I1125 17:19:16.157994 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_19127e7976e8d018e0cb77daad9e54419526b5a7f1c0f33dd6ea52c1ffbn26k_9ad4315c-9d47-485c-b656-4ba2e41803f6/extract/0.log" Nov 25 17:19:16 crc kubenswrapper[4802]: I1125 17:19:16.282695 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn_f6472f81-bf27-4839-a0a4-d7baa206c138/util/0.log" Nov 25 17:19:16 crc kubenswrapper[4802]: I1125 17:19:16.476766 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn_f6472f81-bf27-4839-a0a4-d7baa206c138/pull/0.log" Nov 25 17:19:16 crc kubenswrapper[4802]: I1125 17:19:16.486477 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn_f6472f81-bf27-4839-a0a4-d7baa206c138/pull/0.log" Nov 25 17:19:16 crc kubenswrapper[4802]: I1125 17:19:16.610893 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn_f6472f81-bf27-4839-a0a4-d7baa206c138/util/0.log" Nov 25 17:19:16 crc kubenswrapper[4802]: I1125 17:19:16.750352 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn_f6472f81-bf27-4839-a0a4-d7baa206c138/util/0.log" Nov 25 17:19:16 crc kubenswrapper[4802]: I1125 17:19:16.750992 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn_f6472f81-bf27-4839-a0a4-d7baa206c138/pull/0.log" Nov 25 17:19:16 crc kubenswrapper[4802]: I1125 17:19:16.775378 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d8wqrn_f6472f81-bf27-4839-a0a4-d7baa206c138/extract/0.log" Nov 25 17:19:16 crc kubenswrapper[4802]: I1125 17:19:16.895471 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7_5b0bee27-5168-4984-9943-c93299d3ac58/util/0.log" Nov 25 17:19:17 crc kubenswrapper[4802]: I1125 17:19:17.046422 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7_5b0bee27-5168-4984-9943-c93299d3ac58/util/0.log" Nov 25 17:19:17 crc kubenswrapper[4802]: I1125 17:19:17.070277 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7_5b0bee27-5168-4984-9943-c93299d3ac58/pull/0.log" Nov 25 17:19:17 crc kubenswrapper[4802]: I1125 17:19:17.087097 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7_5b0bee27-5168-4984-9943-c93299d3ac58/pull/0.log" Nov 25 17:19:17 crc kubenswrapper[4802]: I1125 17:19:17.243952 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7_5b0bee27-5168-4984-9943-c93299d3ac58/util/0.log" Nov 25 17:19:17 crc kubenswrapper[4802]: I1125 17:19:17.248374 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7_5b0bee27-5168-4984-9943-c93299d3ac58/pull/0.log" Nov 25 17:19:17 crc kubenswrapper[4802]: I1125 17:19:17.267747 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bb68r7_5b0bee27-5168-4984-9943-c93299d3ac58/extract/0.log" Nov 25 17:19:17 crc kubenswrapper[4802]: I1125 17:19:17.455824 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc_376dff9f-509e-46c3-b110-8bf2c43dffda/util/0.log" Nov 25 17:19:17 crc kubenswrapper[4802]: I1125 17:19:17.627645 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc_376dff9f-509e-46c3-b110-8bf2c43dffda/pull/0.log" Nov 25 17:19:17 crc kubenswrapper[4802]: I1125 17:19:17.631624 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc_376dff9f-509e-46c3-b110-8bf2c43dffda/util/0.log" Nov 25 17:19:17 crc kubenswrapper[4802]: I1125 17:19:17.664394 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc_376dff9f-509e-46c3-b110-8bf2c43dffda/pull/0.log" Nov 25 17:19:17 crc kubenswrapper[4802]: I1125 17:19:17.803549 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc_376dff9f-509e-46c3-b110-8bf2c43dffda/extract/0.log" Nov 25 17:19:17 crc kubenswrapper[4802]: I1125 17:19:17.835883 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc_376dff9f-509e-46c3-b110-8bf2c43dffda/util/0.log" Nov 25 17:19:17 crc kubenswrapper[4802]: I1125 17:19:17.846824 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rfhqc_376dff9f-509e-46c3-b110-8bf2c43dffda/pull/0.log" Nov 25 17:19:17 crc kubenswrapper[4802]: I1125 17:19:17.994597 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr_d0befe17-9e51-457b-86c7-8831cb064339/util/0.log" Nov 25 17:19:18 crc kubenswrapper[4802]: I1125 17:19:18.149747 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr_d0befe17-9e51-457b-86c7-8831cb064339/util/0.log" Nov 25 17:19:18 crc kubenswrapper[4802]: I1125 17:19:18.178054 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr_d0befe17-9e51-457b-86c7-8831cb064339/pull/0.log" Nov 25 17:19:18 crc kubenswrapper[4802]: I1125 17:19:18.210947 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr_d0befe17-9e51-457b-86c7-8831cb064339/pull/0.log" Nov 25 17:19:18 crc kubenswrapper[4802]: I1125 17:19:18.408527 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr_d0befe17-9e51-457b-86c7-8831cb064339/extract/0.log" Nov 25 17:19:18 crc kubenswrapper[4802]: I1125 17:19:18.426030 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr_d0befe17-9e51-457b-86c7-8831cb064339/util/0.log" Nov 25 17:19:18 crc kubenswrapper[4802]: I1125 17:19:18.456951 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c7kpgr_d0befe17-9e51-457b-86c7-8831cb064339/pull/0.log" Nov 25 17:19:18 crc kubenswrapper[4802]: I1125 17:19:18.541858 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc_d3f9606c-d757-40c8-a32e-52b5d9572f65/util/0.log" Nov 25 17:19:18 crc kubenswrapper[4802]: I1125 17:19:18.709236 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc_d3f9606c-d757-40c8-a32e-52b5d9572f65/util/0.log" Nov 25 17:19:18 crc kubenswrapper[4802]: I1125 17:19:18.709434 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc_d3f9606c-d757-40c8-a32e-52b5d9572f65/pull/0.log" Nov 25 17:19:18 crc kubenswrapper[4802]: I1125 17:19:18.715857 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc_d3f9606c-d757-40c8-a32e-52b5d9572f65/pull/0.log" Nov 25 17:19:18 crc kubenswrapper[4802]: I1125 17:19:18.851618 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc_d3f9606c-d757-40c8-a32e-52b5d9572f65/util/0.log" Nov 25 17:19:18 crc kubenswrapper[4802]: I1125 17:19:18.880250 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc_d3f9606c-d757-40c8-a32e-52b5d9572f65/extract/0.log" Nov 25 17:19:18 crc kubenswrapper[4802]: I1125 17:19:18.924886 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f68cxlc_d3f9606c-d757-40c8-a32e-52b5d9572f65/pull/0.log" Nov 25 17:19:18 crc kubenswrapper[4802]: I1125 17:19:18.944263 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l_2e812eef-efff-4d2b-a943-57b1305d73eb/util/0.log" Nov 25 17:19:19 crc kubenswrapper[4802]: I1125 17:19:19.063291 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l_2e812eef-efff-4d2b-a943-57b1305d73eb/util/0.log" Nov 25 17:19:19 crc kubenswrapper[4802]: I1125 17:19:19.069536 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l_2e812eef-efff-4d2b-a943-57b1305d73eb/pull/0.log" Nov 25 17:19:19 crc kubenswrapper[4802]: I1125 17:19:19.096254 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l_2e812eef-efff-4d2b-a943-57b1305d73eb/pull/0.log" Nov 25 17:19:19 crc kubenswrapper[4802]: I1125 17:19:19.231769 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l_2e812eef-efff-4d2b-a943-57b1305d73eb/util/0.log" Nov 25 17:19:19 crc kubenswrapper[4802]: I1125 17:19:19.255317 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l_2e812eef-efff-4d2b-a943-57b1305d73eb/extract/0.log" Nov 25 17:19:19 crc kubenswrapper[4802]: I1125 17:19:19.257342 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3slb5l_2e812eef-efff-4d2b-a943-57b1305d73eb/pull/0.log" Nov 25 17:19:19 crc kubenswrapper[4802]: I1125 17:19:19.305835 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-6bfbf5c856-xw6fj_eeca972a-357e-4fa8-a8b5-1421b6601219/manager/3.log" Nov 25 17:19:19 crc kubenswrapper[4802]: I1125 17:19:19.440240 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-6bfbf5c856-xw6fj_eeca972a-357e-4fa8-a8b5-1421b6601219/manager/2.log" Nov 25 17:19:19 crc kubenswrapper[4802]: I1125 17:19:19.441919 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-index-sdnvr_04dd5ba3-51fb-4edc-9d85-65fde00950de/registry-server/0.log" Nov 25 17:19:19 crc kubenswrapper[4802]: I1125 17:19:19.483357 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-d58bc84f4-fjc7z_c328497f-325f-4500-ba48-d1523cd7ee74/manager/3.log" Nov 25 17:19:19 crc kubenswrapper[4802]: I1125 17:19:19.622630 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-d58bc84f4-fjc7z_c328497f-325f-4500-ba48-d1523cd7ee74/manager/2.log" Nov 25 17:19:19 crc kubenswrapper[4802]: I1125 17:19:19.676182 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-dt6d7_bce25e4a-fd42-4ffd-9c42-fad1e41af5e6/registry-server/0.log" Nov 25 17:19:19 crc kubenswrapper[4802]: I1125 17:19:19.703596 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6fccf5cdb-rrwcf_eda71d4e-8d95-4944-a0b8-a8ac84c01ca4/kube-rbac-proxy/0.log" Nov 25 17:19:19 crc kubenswrapper[4802]: I1125 17:19:19.801763 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6fccf5cdb-rrwcf_eda71d4e-8d95-4944-a0b8-a8ac84c01ca4/manager/3.log" Nov 25 17:19:19 crc kubenswrapper[4802]: I1125 17:19:19.856769 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6fccf5cdb-rrwcf_eda71d4e-8d95-4944-a0b8-a8ac84c01ca4/manager/2.log" Nov 25 17:19:19 crc kubenswrapper[4802]: I1125 17:19:19.891204 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-khg9j_f8f50f91-e0de-42a8-b9b2-4b4eff04bd74/registry-server/0.log" Nov 25 17:19:20 crc kubenswrapper[4802]: I1125 17:19:20.033051 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-65c6fdb94d-5cfbr_a5bdabd0-7540-4d45-8675-9bd88c415957/manager/2.log" Nov 25 17:19:20 crc kubenswrapper[4802]: I1125 17:19:20.039081 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-65c6fdb94d-5cfbr_a5bdabd0-7540-4d45-8675-9bd88c415957/manager/3.log" Nov 25 17:19:20 crc kubenswrapper[4802]: I1125 17:19:20.126692 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-bs696_e9c24dd4-6edb-4e3f-90fd-5a1230fcf452/registry-server/0.log" Nov 25 17:19:20 crc kubenswrapper[4802]: I1125 17:19:20.201269 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5959d99654-4j2qc_5004bc39-3a06-4842-9565-8501f35c8c4b/manager/3.log" Nov 25 17:19:20 crc kubenswrapper[4802]: I1125 17:19:20.210628 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5959d99654-4j2qc_5004bc39-3a06-4842-9565-8501f35c8c4b/manager/2.log" Nov 25 17:19:20 crc kubenswrapper[4802]: I1125 17:19:20.286394 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-index-nwwvv_ad830c47-f0b5-48d7-874f-41f78fbaf23d/registry-server/0.log" Nov 25 17:19:20 crc kubenswrapper[4802]: I1125 17:19:20.370520 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-987nc_373fd8b2-c469-46b8-b20e-8ecd875c4a39/operator/3.log" Nov 25 17:19:20 crc kubenswrapper[4802]: I1125 17:19:20.380920 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-987nc_373fd8b2-c469-46b8-b20e-8ecd875c4a39/operator/2.log" Nov 25 17:19:20 crc kubenswrapper[4802]: I1125 17:19:20.459620 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-index-984qp_7fec300a-5625-4876-b0a0-d82c18409f4b/registry-server/0.log" Nov 25 17:19:20 crc kubenswrapper[4802]: I1125 17:19:20.549966 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5787d846c-g6b4c_3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb/manager/2.log" Nov 25 17:19:20 crc kubenswrapper[4802]: I1125 17:19:20.557986 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5787d846c-g6b4c_3d4897e2-ae62-4c61-88c6-f1f3ff6b6abb/manager/3.log" Nov 25 17:19:20 crc kubenswrapper[4802]: I1125 17:19:20.672337 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-index-hp7k9_459e8530-a5fb-4b88-b42d-b68cf4d41835/registry-server/0.log" Nov 25 17:19:36 crc kubenswrapper[4802]: I1125 17:19:36.505623 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-9lzl9_7583c399-65b2-4e40-91eb-1908753b6a29/control-plane-machine-set-operator/0.log" Nov 25 17:19:36 crc kubenswrapper[4802]: I1125 17:19:36.614313 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j2t9z_ea30ecf0-8806-4c5a-9ff6-2738ca72e50b/kube-rbac-proxy/0.log" Nov 25 17:19:36 crc kubenswrapper[4802]: I1125 17:19:36.658724 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j2t9z_ea30ecf0-8806-4c5a-9ff6-2738ca72e50b/machine-api-operator/0.log" Nov 25 17:19:53 crc kubenswrapper[4802]: I1125 17:19:53.029682 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-tcf29_85895137-3ed8-4214-92fa-ea23980c5083/kube-rbac-proxy/0.log" Nov 25 17:19:53 crc kubenswrapper[4802]: I1125 17:19:53.076972 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-tcf29_85895137-3ed8-4214-92fa-ea23980c5083/controller/0.log" Nov 25 17:19:53 crc kubenswrapper[4802]: I1125 17:19:53.166931 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rzq2t_aee28310-d76a-491b-b174-e1d131d078c7/cp-frr-files/0.log" Nov 25 17:19:53 crc kubenswrapper[4802]: I1125 17:19:53.329401 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rzq2t_aee28310-d76a-491b-b174-e1d131d078c7/cp-frr-files/0.log" Nov 25 17:19:53 crc kubenswrapper[4802]: I1125 17:19:53.355661 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rzq2t_aee28310-d76a-491b-b174-e1d131d078c7/cp-metrics/0.log" Nov 25 17:19:53 crc kubenswrapper[4802]: I1125 17:19:53.395160 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rzq2t_aee28310-d76a-491b-b174-e1d131d078c7/cp-reloader/0.log" Nov 25 17:19:53 crc kubenswrapper[4802]: I1125 17:19:53.425734 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rzq2t_aee28310-d76a-491b-b174-e1d131d078c7/cp-reloader/0.log" Nov 25 17:19:53 crc kubenswrapper[4802]: I1125 17:19:53.533197 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rzq2t_aee28310-d76a-491b-b174-e1d131d078c7/cp-reloader/0.log" Nov 25 17:19:53 crc kubenswrapper[4802]: I1125 17:19:53.549789 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rzq2t_aee28310-d76a-491b-b174-e1d131d078c7/cp-metrics/0.log" Nov 25 17:19:53 crc kubenswrapper[4802]: I1125 17:19:53.567693 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rzq2t_aee28310-d76a-491b-b174-e1d131d078c7/cp-frr-files/0.log" Nov 25 17:19:53 crc kubenswrapper[4802]: I1125 17:19:53.627004 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rzq2t_aee28310-d76a-491b-b174-e1d131d078c7/cp-metrics/0.log" Nov 25 17:19:53 crc kubenswrapper[4802]: I1125 17:19:53.796563 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rzq2t_aee28310-d76a-491b-b174-e1d131d078c7/cp-reloader/0.log" Nov 25 17:19:53 crc kubenswrapper[4802]: I1125 17:19:53.810556 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rzq2t_aee28310-d76a-491b-b174-e1d131d078c7/cp-frr-files/0.log" Nov 25 17:19:53 crc kubenswrapper[4802]: I1125 17:19:53.818211 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rzq2t_aee28310-d76a-491b-b174-e1d131d078c7/cp-metrics/0.log" Nov 25 17:19:53 crc kubenswrapper[4802]: I1125 17:19:53.885555 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rzq2t_aee28310-d76a-491b-b174-e1d131d078c7/controller/0.log" Nov 25 17:19:54 crc kubenswrapper[4802]: I1125 17:19:54.017691 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rzq2t_aee28310-d76a-491b-b174-e1d131d078c7/kube-rbac-proxy/0.log" Nov 25 17:19:54 crc kubenswrapper[4802]: I1125 17:19:54.084194 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rzq2t_aee28310-d76a-491b-b174-e1d131d078c7/frr-metrics/0.log" Nov 25 17:19:54 crc kubenswrapper[4802]: I1125 17:19:54.094363 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rzq2t_aee28310-d76a-491b-b174-e1d131d078c7/kube-rbac-proxy-frr/0.log" Nov 25 17:19:54 crc kubenswrapper[4802]: I1125 17:19:54.247976 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 17:19:54 crc kubenswrapper[4802]: I1125 17:19:54.248298 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 17:19:54 crc kubenswrapper[4802]: I1125 17:19:54.256164 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rzq2t_aee28310-d76a-491b-b174-e1d131d078c7/reloader/0.log" Nov 25 17:19:54 crc kubenswrapper[4802]: I1125 17:19:54.280986 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-rzknk_d605e3c4-111d-4f35-bf63-f855001d1967/frr-k8s-webhook-server/0.log" Nov 25 17:19:54 crc kubenswrapper[4802]: I1125 17:19:54.444808 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rzq2t_aee28310-d76a-491b-b174-e1d131d078c7/frr/0.log" Nov 25 17:19:54 crc kubenswrapper[4802]: I1125 17:19:54.521659 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-548f76d7c-s5g2f_a1724655-9ac1-43dc-8292-f39870b4a855/manager/3.log" Nov 25 17:19:54 crc kubenswrapper[4802]: I1125 17:19:54.541866 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-548f76d7c-s5g2f_a1724655-9ac1-43dc-8292-f39870b4a855/manager/2.log" Nov 25 17:19:54 crc kubenswrapper[4802]: I1125 17:19:54.691900 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-867dbdb569-pqqsd_77ecf1f1-1f2d-49a6-af38-32d16810ae9d/webhook-server/0.log" Nov 25 17:19:54 crc kubenswrapper[4802]: I1125 17:19:54.708272 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-xnshj_cd0861df-163a-47f7-8a35-8b8c25e02ca4/kube-rbac-proxy/0.log" Nov 25 17:19:54 crc kubenswrapper[4802]: I1125 17:19:54.965710 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-xnshj_cd0861df-163a-47f7-8a35-8b8c25e02ca4/speaker/0.log" Nov 25 17:20:08 crc kubenswrapper[4802]: I1125 17:20:08.486169 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_keystone-64bc67949c-jgvwk_b799cfff-4f77-4b6c-928e-0bcbed6fc523/keystone-api/0.log" Nov 25 17:20:08 crc kubenswrapper[4802]: I1125 17:20:08.495803 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_6d23bfa4-e3e1-4942-82d3-dc6d548b0a98/mysql-bootstrap/0.log" Nov 25 17:20:08 crc kubenswrapper[4802]: I1125 17:20:08.708951 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_6d23bfa4-e3e1-4942-82d3-dc6d548b0a98/mysql-bootstrap/0.log" Nov 25 17:20:08 crc kubenswrapper[4802]: I1125 17:20:08.777470 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_6d23bfa4-e3e1-4942-82d3-dc6d548b0a98/galera/0.log" Nov 25 17:20:08 crc kubenswrapper[4802]: I1125 17:20:08.940670 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab/mysql-bootstrap/0.log" Nov 25 17:20:09 crc kubenswrapper[4802]: I1125 17:20:09.154506 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab/mysql-bootstrap/0.log" Nov 25 17:20:09 crc kubenswrapper[4802]: I1125 17:20:09.157563 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_4f15a614-5ba1-42d4-a9d5-1daa1a7bf8ab/galera/0.log" Nov 25 17:20:09 crc kubenswrapper[4802]: I1125 17:20:09.342646 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_3ea5778f-cef3-42cd-b4bb-1fc73a0ee456/mysql-bootstrap/0.log" Nov 25 17:20:09 crc kubenswrapper[4802]: I1125 17:20:09.510450 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_3ea5778f-cef3-42cd-b4bb-1fc73a0ee456/mysql-bootstrap/0.log" Nov 25 17:20:09 crc kubenswrapper[4802]: I1125 17:20:09.524372 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_3ea5778f-cef3-42cd-b4bb-1fc73a0ee456/galera/0.log" Nov 25 17:20:09 crc kubenswrapper[4802]: I1125 17:20:09.703205 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_5f67491f-4728-432f-ae26-a82724aade76/openstackclient/0.log" Nov 25 17:20:09 crc kubenswrapper[4802]: I1125 17:20:09.768849 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_74e76589-71a7-4c7f-bbac-edee84556bd7/setup-container/0.log" Nov 25 17:20:09 crc kubenswrapper[4802]: I1125 17:20:09.929318 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_74e76589-71a7-4c7f-bbac-edee84556bd7/setup-container/0.log" Nov 25 17:20:09 crc kubenswrapper[4802]: I1125 17:20:09.988720 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_74e76589-71a7-4c7f-bbac-edee84556bd7/rabbitmq/0.log" Nov 25 17:20:10 crc kubenswrapper[4802]: I1125 17:20:10.091076 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_memcached-0_8357d6cc-810d-42bc-b387-f3972fede479/memcached/0.log" Nov 25 17:20:10 crc kubenswrapper[4802]: I1125 17:20:10.136962 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-6bd58cfcf7-4q4kb_b82f4893-deaf-4760-8448-30dbc4c5e100/proxy-httpd/0.log" Nov 25 17:20:10 crc kubenswrapper[4802]: I1125 17:20:10.181691 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-6bd58cfcf7-4q4kb_b82f4893-deaf-4760-8448-30dbc4c5e100/proxy-server/0.log" Nov 25 17:20:10 crc kubenswrapper[4802]: I1125 17:20:10.238581 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-ring-rebalance-m4h5c_305295e6-42a5-41a9-ad55-17945c616b87/swift-ring-rebalance/0.log" Nov 25 17:20:10 crc kubenswrapper[4802]: I1125 17:20:10.398831 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_4a121db3-b69c-484d-9917-f13c0c8ac50f/account-auditor/0.log" Nov 25 17:20:10 crc kubenswrapper[4802]: I1125 17:20:10.432401 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_4a121db3-b69c-484d-9917-f13c0c8ac50f/account-reaper/0.log" Nov 25 17:20:10 crc kubenswrapper[4802]: I1125 17:20:10.457628 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_4a121db3-b69c-484d-9917-f13c0c8ac50f/account-replicator/0.log" Nov 25 17:20:10 crc kubenswrapper[4802]: I1125 17:20:10.504324 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_4a121db3-b69c-484d-9917-f13c0c8ac50f/account-server/0.log" Nov 25 17:20:10 crc kubenswrapper[4802]: I1125 17:20:10.598704 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_4a121db3-b69c-484d-9917-f13c0c8ac50f/container-auditor/0.log" Nov 25 17:20:10 crc kubenswrapper[4802]: I1125 17:20:10.619081 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_4a121db3-b69c-484d-9917-f13c0c8ac50f/container-replicator/0.log" Nov 25 17:20:10 crc kubenswrapper[4802]: I1125 17:20:10.656439 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_4a121db3-b69c-484d-9917-f13c0c8ac50f/container-server/0.log" Nov 25 17:20:10 crc kubenswrapper[4802]: I1125 17:20:10.741589 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_4a121db3-b69c-484d-9917-f13c0c8ac50f/container-updater/0.log" Nov 25 17:20:10 crc kubenswrapper[4802]: I1125 17:20:10.873208 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_4a121db3-b69c-484d-9917-f13c0c8ac50f/object-auditor/0.log" Nov 25 17:20:10 crc kubenswrapper[4802]: I1125 17:20:10.966161 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_4a121db3-b69c-484d-9917-f13c0c8ac50f/object-replicator/0.log" Nov 25 17:20:10 crc kubenswrapper[4802]: I1125 17:20:10.975026 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_4a121db3-b69c-484d-9917-f13c0c8ac50f/object-expirer/0.log" Nov 25 17:20:10 crc kubenswrapper[4802]: I1125 17:20:10.993036 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_4a121db3-b69c-484d-9917-f13c0c8ac50f/object-server/0.log" Nov 25 17:20:11 crc kubenswrapper[4802]: I1125 17:20:11.060907 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_4a121db3-b69c-484d-9917-f13c0c8ac50f/object-updater/0.log" Nov 25 17:20:11 crc kubenswrapper[4802]: I1125 17:20:11.127561 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_4a121db3-b69c-484d-9917-f13c0c8ac50f/rsync/0.log" Nov 25 17:20:11 crc kubenswrapper[4802]: I1125 17:20:11.165807 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_4a121db3-b69c-484d-9917-f13c0c8ac50f/swift-recon-cron/0.log" Nov 25 17:20:24 crc kubenswrapper[4802]: I1125 17:20:24.248605 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 17:20:24 crc kubenswrapper[4802]: I1125 17:20:24.249137 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 17:20:24 crc kubenswrapper[4802]: I1125 17:20:24.498641 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rslrk_6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd/extract-utilities/0.log" Nov 25 17:20:24 crc kubenswrapper[4802]: I1125 17:20:24.665715 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rslrk_6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd/extract-utilities/0.log" Nov 25 17:20:24 crc kubenswrapper[4802]: I1125 17:20:24.691545 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rslrk_6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd/extract-content/0.log" Nov 25 17:20:24 crc kubenswrapper[4802]: I1125 17:20:24.730020 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rslrk_6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd/extract-content/0.log" Nov 25 17:20:24 crc kubenswrapper[4802]: I1125 17:20:24.873662 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rslrk_6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd/extract-utilities/0.log" Nov 25 17:20:24 crc kubenswrapper[4802]: I1125 17:20:24.956164 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rslrk_6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd/extract-content/0.log" Nov 25 17:20:25 crc kubenswrapper[4802]: I1125 17:20:25.060903 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mlz5g_ba359221-8966-436b-b846-4c5ac2f1767e/extract-utilities/0.log" Nov 25 17:20:25 crc kubenswrapper[4802]: I1125 17:20:25.397736 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mlz5g_ba359221-8966-436b-b846-4c5ac2f1767e/extract-utilities/0.log" Nov 25 17:20:25 crc kubenswrapper[4802]: I1125 17:20:25.451824 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mlz5g_ba359221-8966-436b-b846-4c5ac2f1767e/extract-content/0.log" Nov 25 17:20:25 crc kubenswrapper[4802]: I1125 17:20:25.468369 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mlz5g_ba359221-8966-436b-b846-4c5ac2f1767e/extract-content/0.log" Nov 25 17:20:25 crc kubenswrapper[4802]: I1125 17:20:25.546544 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rslrk_6afc1234-7d7b-4da1-a2c3-62ee8e50b2bd/registry-server/0.log" Nov 25 17:20:25 crc kubenswrapper[4802]: I1125 17:20:25.611012 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mlz5g_ba359221-8966-436b-b846-4c5ac2f1767e/extract-utilities/0.log" Nov 25 17:20:25 crc kubenswrapper[4802]: I1125 17:20:25.649909 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mlz5g_ba359221-8966-436b-b846-4c5ac2f1767e/extract-content/0.log" Nov 25 17:20:25 crc kubenswrapper[4802]: I1125 17:20:25.805785 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_2ef5d76c-957b-4785-8165-50d2b55790c7/util/0.log" Nov 25 17:20:26 crc kubenswrapper[4802]: I1125 17:20:26.020072 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_2ef5d76c-957b-4785-8165-50d2b55790c7/pull/0.log" Nov 25 17:20:26 crc kubenswrapper[4802]: I1125 17:20:26.024353 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_2ef5d76c-957b-4785-8165-50d2b55790c7/pull/0.log" Nov 25 17:20:26 crc kubenswrapper[4802]: I1125 17:20:26.091336 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_2ef5d76c-957b-4785-8165-50d2b55790c7/util/0.log" Nov 25 17:20:26 crc kubenswrapper[4802]: I1125 17:20:26.241136 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_2ef5d76c-957b-4785-8165-50d2b55790c7/util/0.log" Nov 25 17:20:26 crc kubenswrapper[4802]: I1125 17:20:26.330503 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_2ef5d76c-957b-4785-8165-50d2b55790c7/extract/0.log" Nov 25 17:20:26 crc kubenswrapper[4802]: I1125 17:20:26.345314 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mlz5g_ba359221-8966-436b-b846-4c5ac2f1767e/registry-server/0.log" Nov 25 17:20:26 crc kubenswrapper[4802]: I1125 17:20:26.348044 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64kdg4_2ef5d76c-957b-4785-8165-50d2b55790c7/pull/0.log" Nov 25 17:20:26 crc kubenswrapper[4802]: I1125 17:20:26.532879 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-rxhj4_90e4d7a1-f42a-4a46-9dc7-a8778f357bb7/marketplace-operator/0.log" Nov 25 17:20:26 crc kubenswrapper[4802]: I1125 17:20:26.542865 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f5llz_b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5/extract-utilities/0.log" Nov 25 17:20:26 crc kubenswrapper[4802]: I1125 17:20:26.688989 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f5llz_b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5/extract-utilities/0.log" Nov 25 17:20:26 crc kubenswrapper[4802]: I1125 17:20:26.724203 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f5llz_b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5/extract-content/0.log" Nov 25 17:20:26 crc kubenswrapper[4802]: I1125 17:20:26.738473 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f5llz_b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5/extract-content/0.log" Nov 25 17:20:26 crc kubenswrapper[4802]: I1125 17:20:26.926403 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f5llz_b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5/extract-content/0.log" Nov 25 17:20:26 crc kubenswrapper[4802]: I1125 17:20:26.927382 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f5llz_b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5/extract-utilities/0.log" Nov 25 17:20:27 crc kubenswrapper[4802]: I1125 17:20:27.064449 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f5llz_b2ae61bd-15c2-47c4-9cf0-8161fae2f3d5/registry-server/0.log" Nov 25 17:20:27 crc kubenswrapper[4802]: I1125 17:20:27.143675 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kw428_495e0a1d-2dd0-4069-a9dd-6637b835a35a/extract-utilities/0.log" Nov 25 17:20:27 crc kubenswrapper[4802]: I1125 17:20:27.293680 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kw428_495e0a1d-2dd0-4069-a9dd-6637b835a35a/extract-content/0.log" Nov 25 17:20:27 crc kubenswrapper[4802]: I1125 17:20:27.308404 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kw428_495e0a1d-2dd0-4069-a9dd-6637b835a35a/extract-content/0.log" Nov 25 17:20:27 crc kubenswrapper[4802]: I1125 17:20:27.329361 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kw428_495e0a1d-2dd0-4069-a9dd-6637b835a35a/extract-utilities/0.log" Nov 25 17:20:27 crc kubenswrapper[4802]: I1125 17:20:27.473462 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kw428_495e0a1d-2dd0-4069-a9dd-6637b835a35a/extract-utilities/0.log" Nov 25 17:20:27 crc kubenswrapper[4802]: I1125 17:20:27.524612 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kw428_495e0a1d-2dd0-4069-a9dd-6637b835a35a/extract-content/0.log" Nov 25 17:20:27 crc kubenswrapper[4802]: I1125 17:20:27.988652 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kw428_495e0a1d-2dd0-4069-a9dd-6637b835a35a/registry-server/0.log" Nov 25 17:20:54 crc kubenswrapper[4802]: I1125 17:20:54.248304 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 17:20:54 crc kubenswrapper[4802]: I1125 17:20:54.248857 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 17:20:54 crc kubenswrapper[4802]: I1125 17:20:54.248913 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 17:20:54 crc kubenswrapper[4802]: I1125 17:20:54.250219 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0ba8f23d451c7d83e7568a15af490bcf0f8abaf83f184136081d26b773dddc2e"} pod="openshift-machine-config-operator/machine-config-daemon-h29wc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 17:20:54 crc kubenswrapper[4802]: I1125 17:20:54.250326 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" containerID="cri-o://0ba8f23d451c7d83e7568a15af490bcf0f8abaf83f184136081d26b773dddc2e" gracePeriod=600 Nov 25 17:20:55 crc kubenswrapper[4802]: I1125 17:20:55.156992 4802 generic.go:334] "Generic (PLEG): container finished" podID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerID="0ba8f23d451c7d83e7568a15af490bcf0f8abaf83f184136081d26b773dddc2e" exitCode=0 Nov 25 17:20:55 crc kubenswrapper[4802]: I1125 17:20:55.157193 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerDied","Data":"0ba8f23d451c7d83e7568a15af490bcf0f8abaf83f184136081d26b773dddc2e"} Nov 25 17:20:55 crc kubenswrapper[4802]: I1125 17:20:55.157596 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerStarted","Data":"1535d557614337c68c32573c76f4ce74b2b995bdb4dc21eb0f632bc8caf7401a"} Nov 25 17:20:55 crc kubenswrapper[4802]: I1125 17:20:55.157626 4802 scope.go:117] "RemoveContainer" containerID="28d4e3cbe0830d69a46ab494e4ad85c068421ad9c9741eb17454d5ad5149dd29" Nov 25 17:21:34 crc kubenswrapper[4802]: I1125 17:21:34.535968 4802 generic.go:334] "Generic (PLEG): container finished" podID="a2c3e739-8ad9-4d34-a24e-8f4b204d05f1" containerID="acea1ffae229b7bf64717bebc1f977aa771ae0ce5cf2ba32e01730de226bacd3" exitCode=0 Nov 25 17:21:34 crc kubenswrapper[4802]: I1125 17:21:34.536084 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nn27p/must-gather-spl62" event={"ID":"a2c3e739-8ad9-4d34-a24e-8f4b204d05f1","Type":"ContainerDied","Data":"acea1ffae229b7bf64717bebc1f977aa771ae0ce5cf2ba32e01730de226bacd3"} Nov 25 17:21:34 crc kubenswrapper[4802]: I1125 17:21:34.536938 4802 scope.go:117] "RemoveContainer" containerID="acea1ffae229b7bf64717bebc1f977aa771ae0ce5cf2ba32e01730de226bacd3" Nov 25 17:21:34 crc kubenswrapper[4802]: I1125 17:21:34.666620 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nn27p_must-gather-spl62_a2c3e739-8ad9-4d34-a24e-8f4b204d05f1/gather/0.log" Nov 25 17:21:41 crc kubenswrapper[4802]: I1125 17:21:41.983653 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-nn27p/must-gather-spl62"] Nov 25 17:21:41 crc kubenswrapper[4802]: I1125 17:21:41.984528 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-nn27p/must-gather-spl62" podUID="a2c3e739-8ad9-4d34-a24e-8f4b204d05f1" containerName="copy" containerID="cri-o://c9716fe7de23debe02bc930037841704eaad1522e9e883d62c7fb56ed2ee1dbd" gracePeriod=2 Nov 25 17:21:41 crc kubenswrapper[4802]: I1125 17:21:41.991665 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-nn27p/must-gather-spl62"] Nov 25 17:21:42 crc kubenswrapper[4802]: I1125 17:21:42.383274 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nn27p_must-gather-spl62_a2c3e739-8ad9-4d34-a24e-8f4b204d05f1/copy/0.log" Nov 25 17:21:42 crc kubenswrapper[4802]: I1125 17:21:42.385583 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nn27p/must-gather-spl62" Nov 25 17:21:42 crc kubenswrapper[4802]: I1125 17:21:42.445583 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a2c3e739-8ad9-4d34-a24e-8f4b204d05f1-must-gather-output\") pod \"a2c3e739-8ad9-4d34-a24e-8f4b204d05f1\" (UID: \"a2c3e739-8ad9-4d34-a24e-8f4b204d05f1\") " Nov 25 17:21:42 crc kubenswrapper[4802]: I1125 17:21:42.445694 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmrjb\" (UniqueName: \"kubernetes.io/projected/a2c3e739-8ad9-4d34-a24e-8f4b204d05f1-kube-api-access-kmrjb\") pod \"a2c3e739-8ad9-4d34-a24e-8f4b204d05f1\" (UID: \"a2c3e739-8ad9-4d34-a24e-8f4b204d05f1\") " Nov 25 17:21:42 crc kubenswrapper[4802]: I1125 17:21:42.450257 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2c3e739-8ad9-4d34-a24e-8f4b204d05f1-kube-api-access-kmrjb" (OuterVolumeSpecName: "kube-api-access-kmrjb") pod "a2c3e739-8ad9-4d34-a24e-8f4b204d05f1" (UID: "a2c3e739-8ad9-4d34-a24e-8f4b204d05f1"). InnerVolumeSpecName "kube-api-access-kmrjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:21:42 crc kubenswrapper[4802]: I1125 17:21:42.522999 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2c3e739-8ad9-4d34-a24e-8f4b204d05f1-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "a2c3e739-8ad9-4d34-a24e-8f4b204d05f1" (UID: "a2c3e739-8ad9-4d34-a24e-8f4b204d05f1"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:21:42 crc kubenswrapper[4802]: I1125 17:21:42.548036 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmrjb\" (UniqueName: \"kubernetes.io/projected/a2c3e739-8ad9-4d34-a24e-8f4b204d05f1-kube-api-access-kmrjb\") on node \"crc\" DevicePath \"\"" Nov 25 17:21:42 crc kubenswrapper[4802]: I1125 17:21:42.548079 4802 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a2c3e739-8ad9-4d34-a24e-8f4b204d05f1-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 25 17:21:42 crc kubenswrapper[4802]: I1125 17:21:42.596836 4802 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nn27p_must-gather-spl62_a2c3e739-8ad9-4d34-a24e-8f4b204d05f1/copy/0.log" Nov 25 17:21:42 crc kubenswrapper[4802]: I1125 17:21:42.597281 4802 generic.go:334] "Generic (PLEG): container finished" podID="a2c3e739-8ad9-4d34-a24e-8f4b204d05f1" containerID="c9716fe7de23debe02bc930037841704eaad1522e9e883d62c7fb56ed2ee1dbd" exitCode=143 Nov 25 17:21:42 crc kubenswrapper[4802]: I1125 17:21:42.597366 4802 scope.go:117] "RemoveContainer" containerID="c9716fe7de23debe02bc930037841704eaad1522e9e883d62c7fb56ed2ee1dbd" Nov 25 17:21:42 crc kubenswrapper[4802]: I1125 17:21:42.597431 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nn27p/must-gather-spl62" Nov 25 17:21:42 crc kubenswrapper[4802]: I1125 17:21:42.617862 4802 scope.go:117] "RemoveContainer" containerID="acea1ffae229b7bf64717bebc1f977aa771ae0ce5cf2ba32e01730de226bacd3" Nov 25 17:21:42 crc kubenswrapper[4802]: I1125 17:21:42.660891 4802 scope.go:117] "RemoveContainer" containerID="c9716fe7de23debe02bc930037841704eaad1522e9e883d62c7fb56ed2ee1dbd" Nov 25 17:21:42 crc kubenswrapper[4802]: E1125 17:21:42.661357 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9716fe7de23debe02bc930037841704eaad1522e9e883d62c7fb56ed2ee1dbd\": container with ID starting with c9716fe7de23debe02bc930037841704eaad1522e9e883d62c7fb56ed2ee1dbd not found: ID does not exist" containerID="c9716fe7de23debe02bc930037841704eaad1522e9e883d62c7fb56ed2ee1dbd" Nov 25 17:21:42 crc kubenswrapper[4802]: I1125 17:21:42.661399 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9716fe7de23debe02bc930037841704eaad1522e9e883d62c7fb56ed2ee1dbd"} err="failed to get container status \"c9716fe7de23debe02bc930037841704eaad1522e9e883d62c7fb56ed2ee1dbd\": rpc error: code = NotFound desc = could not find container \"c9716fe7de23debe02bc930037841704eaad1522e9e883d62c7fb56ed2ee1dbd\": container with ID starting with c9716fe7de23debe02bc930037841704eaad1522e9e883d62c7fb56ed2ee1dbd not found: ID does not exist" Nov 25 17:21:42 crc kubenswrapper[4802]: I1125 17:21:42.661425 4802 scope.go:117] "RemoveContainer" containerID="acea1ffae229b7bf64717bebc1f977aa771ae0ce5cf2ba32e01730de226bacd3" Nov 25 17:21:42 crc kubenswrapper[4802]: E1125 17:21:42.661814 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acea1ffae229b7bf64717bebc1f977aa771ae0ce5cf2ba32e01730de226bacd3\": container with ID starting with acea1ffae229b7bf64717bebc1f977aa771ae0ce5cf2ba32e01730de226bacd3 not found: ID does not exist" containerID="acea1ffae229b7bf64717bebc1f977aa771ae0ce5cf2ba32e01730de226bacd3" Nov 25 17:21:42 crc kubenswrapper[4802]: I1125 17:21:42.661850 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acea1ffae229b7bf64717bebc1f977aa771ae0ce5cf2ba32e01730de226bacd3"} err="failed to get container status \"acea1ffae229b7bf64717bebc1f977aa771ae0ce5cf2ba32e01730de226bacd3\": rpc error: code = NotFound desc = could not find container \"acea1ffae229b7bf64717bebc1f977aa771ae0ce5cf2ba32e01730de226bacd3\": container with ID starting with acea1ffae229b7bf64717bebc1f977aa771ae0ce5cf2ba32e01730de226bacd3 not found: ID does not exist" Nov 25 17:21:43 crc kubenswrapper[4802]: I1125 17:21:43.512190 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2c3e739-8ad9-4d34-a24e-8f4b204d05f1" path="/var/lib/kubelet/pods/a2c3e739-8ad9-4d34-a24e-8f4b204d05f1/volumes" Nov 25 17:22:09 crc kubenswrapper[4802]: I1125 17:22:09.029726 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xxb7x"] Nov 25 17:22:09 crc kubenswrapper[4802]: E1125 17:22:09.032486 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2c3e739-8ad9-4d34-a24e-8f4b204d05f1" containerName="gather" Nov 25 17:22:09 crc kubenswrapper[4802]: I1125 17:22:09.032616 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2c3e739-8ad9-4d34-a24e-8f4b204d05f1" containerName="gather" Nov 25 17:22:09 crc kubenswrapper[4802]: E1125 17:22:09.032714 4802 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2c3e739-8ad9-4d34-a24e-8f4b204d05f1" containerName="copy" Nov 25 17:22:09 crc kubenswrapper[4802]: I1125 17:22:09.032791 4802 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2c3e739-8ad9-4d34-a24e-8f4b204d05f1" containerName="copy" Nov 25 17:22:09 crc kubenswrapper[4802]: I1125 17:22:09.033043 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2c3e739-8ad9-4d34-a24e-8f4b204d05f1" containerName="gather" Nov 25 17:22:09 crc kubenswrapper[4802]: I1125 17:22:09.033179 4802 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2c3e739-8ad9-4d34-a24e-8f4b204d05f1" containerName="copy" Nov 25 17:22:09 crc kubenswrapper[4802]: I1125 17:22:09.043839 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xxb7x"] Nov 25 17:22:09 crc kubenswrapper[4802]: I1125 17:22:09.044052 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xxb7x" Nov 25 17:22:09 crc kubenswrapper[4802]: I1125 17:22:09.164325 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/012112b2-13b7-42f2-b249-b443ca36abd8-utilities\") pod \"certified-operators-xxb7x\" (UID: \"012112b2-13b7-42f2-b249-b443ca36abd8\") " pod="openshift-marketplace/certified-operators-xxb7x" Nov 25 17:22:09 crc kubenswrapper[4802]: I1125 17:22:09.164398 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fjhk\" (UniqueName: \"kubernetes.io/projected/012112b2-13b7-42f2-b249-b443ca36abd8-kube-api-access-7fjhk\") pod \"certified-operators-xxb7x\" (UID: \"012112b2-13b7-42f2-b249-b443ca36abd8\") " pod="openshift-marketplace/certified-operators-xxb7x" Nov 25 17:22:09 crc kubenswrapper[4802]: I1125 17:22:09.164491 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/012112b2-13b7-42f2-b249-b443ca36abd8-catalog-content\") pod \"certified-operators-xxb7x\" (UID: \"012112b2-13b7-42f2-b249-b443ca36abd8\") " pod="openshift-marketplace/certified-operators-xxb7x" Nov 25 17:22:09 crc kubenswrapper[4802]: I1125 17:22:09.266351 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/012112b2-13b7-42f2-b249-b443ca36abd8-utilities\") pod \"certified-operators-xxb7x\" (UID: \"012112b2-13b7-42f2-b249-b443ca36abd8\") " pod="openshift-marketplace/certified-operators-xxb7x" Nov 25 17:22:09 crc kubenswrapper[4802]: I1125 17:22:09.266642 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fjhk\" (UniqueName: \"kubernetes.io/projected/012112b2-13b7-42f2-b249-b443ca36abd8-kube-api-access-7fjhk\") pod \"certified-operators-xxb7x\" (UID: \"012112b2-13b7-42f2-b249-b443ca36abd8\") " pod="openshift-marketplace/certified-operators-xxb7x" Nov 25 17:22:09 crc kubenswrapper[4802]: I1125 17:22:09.266776 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/012112b2-13b7-42f2-b249-b443ca36abd8-catalog-content\") pod \"certified-operators-xxb7x\" (UID: \"012112b2-13b7-42f2-b249-b443ca36abd8\") " pod="openshift-marketplace/certified-operators-xxb7x" Nov 25 17:22:09 crc kubenswrapper[4802]: I1125 17:22:09.267098 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/012112b2-13b7-42f2-b249-b443ca36abd8-utilities\") pod \"certified-operators-xxb7x\" (UID: \"012112b2-13b7-42f2-b249-b443ca36abd8\") " pod="openshift-marketplace/certified-operators-xxb7x" Nov 25 17:22:09 crc kubenswrapper[4802]: I1125 17:22:09.267342 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/012112b2-13b7-42f2-b249-b443ca36abd8-catalog-content\") pod \"certified-operators-xxb7x\" (UID: \"012112b2-13b7-42f2-b249-b443ca36abd8\") " pod="openshift-marketplace/certified-operators-xxb7x" Nov 25 17:22:09 crc kubenswrapper[4802]: I1125 17:22:09.294403 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fjhk\" (UniqueName: \"kubernetes.io/projected/012112b2-13b7-42f2-b249-b443ca36abd8-kube-api-access-7fjhk\") pod \"certified-operators-xxb7x\" (UID: \"012112b2-13b7-42f2-b249-b443ca36abd8\") " pod="openshift-marketplace/certified-operators-xxb7x" Nov 25 17:22:09 crc kubenswrapper[4802]: I1125 17:22:09.376186 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xxb7x" Nov 25 17:22:09 crc kubenswrapper[4802]: I1125 17:22:09.871770 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xxb7x"] Nov 25 17:22:10 crc kubenswrapper[4802]: I1125 17:22:10.825409 4802 generic.go:334] "Generic (PLEG): container finished" podID="012112b2-13b7-42f2-b249-b443ca36abd8" containerID="a855df6f3ca9f301ee0acbf65dff75e576f9cf98078289093120ce212cde22a1" exitCode=0 Nov 25 17:22:10 crc kubenswrapper[4802]: I1125 17:22:10.825477 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxb7x" event={"ID":"012112b2-13b7-42f2-b249-b443ca36abd8","Type":"ContainerDied","Data":"a855df6f3ca9f301ee0acbf65dff75e576f9cf98078289093120ce212cde22a1"} Nov 25 17:22:10 crc kubenswrapper[4802]: I1125 17:22:10.826166 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxb7x" event={"ID":"012112b2-13b7-42f2-b249-b443ca36abd8","Type":"ContainerStarted","Data":"f19a5f73cd1905546dca9e0aac63038940a17e6d2347f02dc271c75db1e50f4c"} Nov 25 17:22:10 crc kubenswrapper[4802]: I1125 17:22:10.827738 4802 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 17:22:11 crc kubenswrapper[4802]: I1125 17:22:11.834391 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxb7x" event={"ID":"012112b2-13b7-42f2-b249-b443ca36abd8","Type":"ContainerStarted","Data":"c8d83bf2127a2bbb9a55e5a0179b70eecd7639615fa719c0a066a8857abdb2eb"} Nov 25 17:22:12 crc kubenswrapper[4802]: I1125 17:22:12.844106 4802 generic.go:334] "Generic (PLEG): container finished" podID="012112b2-13b7-42f2-b249-b443ca36abd8" containerID="c8d83bf2127a2bbb9a55e5a0179b70eecd7639615fa719c0a066a8857abdb2eb" exitCode=0 Nov 25 17:22:12 crc kubenswrapper[4802]: I1125 17:22:12.844162 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxb7x" event={"ID":"012112b2-13b7-42f2-b249-b443ca36abd8","Type":"ContainerDied","Data":"c8d83bf2127a2bbb9a55e5a0179b70eecd7639615fa719c0a066a8857abdb2eb"} Nov 25 17:22:13 crc kubenswrapper[4802]: I1125 17:22:13.191166 4802 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zfnrs"] Nov 25 17:22:13 crc kubenswrapper[4802]: I1125 17:22:13.192786 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zfnrs" Nov 25 17:22:13 crc kubenswrapper[4802]: I1125 17:22:13.204741 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zfnrs"] Nov 25 17:22:13 crc kubenswrapper[4802]: I1125 17:22:13.247424 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9-utilities\") pod \"redhat-operators-zfnrs\" (UID: \"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9\") " pod="openshift-marketplace/redhat-operators-zfnrs" Nov 25 17:22:13 crc kubenswrapper[4802]: I1125 17:22:13.247628 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9-catalog-content\") pod \"redhat-operators-zfnrs\" (UID: \"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9\") " pod="openshift-marketplace/redhat-operators-zfnrs" Nov 25 17:22:13 crc kubenswrapper[4802]: I1125 17:22:13.247717 4802 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnfld\" (UniqueName: \"kubernetes.io/projected/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9-kube-api-access-xnfld\") pod \"redhat-operators-zfnrs\" (UID: \"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9\") " pod="openshift-marketplace/redhat-operators-zfnrs" Nov 25 17:22:13 crc kubenswrapper[4802]: I1125 17:22:13.349884 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnfld\" (UniqueName: \"kubernetes.io/projected/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9-kube-api-access-xnfld\") pod \"redhat-operators-zfnrs\" (UID: \"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9\") " pod="openshift-marketplace/redhat-operators-zfnrs" Nov 25 17:22:13 crc kubenswrapper[4802]: I1125 17:22:13.350015 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9-utilities\") pod \"redhat-operators-zfnrs\" (UID: \"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9\") " pod="openshift-marketplace/redhat-operators-zfnrs" Nov 25 17:22:13 crc kubenswrapper[4802]: I1125 17:22:13.350052 4802 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9-catalog-content\") pod \"redhat-operators-zfnrs\" (UID: \"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9\") " pod="openshift-marketplace/redhat-operators-zfnrs" Nov 25 17:22:13 crc kubenswrapper[4802]: I1125 17:22:13.350455 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9-catalog-content\") pod \"redhat-operators-zfnrs\" (UID: \"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9\") " pod="openshift-marketplace/redhat-operators-zfnrs" Nov 25 17:22:13 crc kubenswrapper[4802]: I1125 17:22:13.350533 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9-utilities\") pod \"redhat-operators-zfnrs\" (UID: \"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9\") " pod="openshift-marketplace/redhat-operators-zfnrs" Nov 25 17:22:13 crc kubenswrapper[4802]: I1125 17:22:13.372247 4802 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnfld\" (UniqueName: \"kubernetes.io/projected/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9-kube-api-access-xnfld\") pod \"redhat-operators-zfnrs\" (UID: \"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9\") " pod="openshift-marketplace/redhat-operators-zfnrs" Nov 25 17:22:13 crc kubenswrapper[4802]: I1125 17:22:13.557312 4802 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zfnrs" Nov 25 17:22:13 crc kubenswrapper[4802]: I1125 17:22:13.777092 4802 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zfnrs"] Nov 25 17:22:13 crc kubenswrapper[4802]: I1125 17:22:13.856455 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxb7x" event={"ID":"012112b2-13b7-42f2-b249-b443ca36abd8","Type":"ContainerStarted","Data":"82f089cd120da02f777e5947320755ed7736dcb2e09485c465859e2eb79ce353"} Nov 25 17:22:13 crc kubenswrapper[4802]: I1125 17:22:13.857380 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfnrs" event={"ID":"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9","Type":"ContainerStarted","Data":"fbf99ba58a86850d12d2ba889a6d0b9311d25a0777920ff147ddf7c5a7bb23d0"} Nov 25 17:22:13 crc kubenswrapper[4802]: I1125 17:22:13.875654 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xxb7x" podStartSLOduration=3.417075124 podStartE2EDuration="5.875637835s" podCreationTimestamp="2025-11-25 17:22:08 +0000 UTC" firstStartedPulling="2025-11-25 17:22:10.827487304 +0000 UTC m=+2113.971834490" lastFinishedPulling="2025-11-25 17:22:13.286050015 +0000 UTC m=+2116.430397201" observedRunningTime="2025-11-25 17:22:13.870409724 +0000 UTC m=+2117.014756910" watchObservedRunningTime="2025-11-25 17:22:13.875637835 +0000 UTC m=+2117.019985021" Nov 25 17:22:14 crc kubenswrapper[4802]: I1125 17:22:14.868305 4802 generic.go:334] "Generic (PLEG): container finished" podID="c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9" containerID="b01a93e161f49bd09c9973506ab5fe658bf6cb6f051521206b34253a4b3a90c4" exitCode=0 Nov 25 17:22:14 crc kubenswrapper[4802]: I1125 17:22:14.868405 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfnrs" event={"ID":"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9","Type":"ContainerDied","Data":"b01a93e161f49bd09c9973506ab5fe658bf6cb6f051521206b34253a4b3a90c4"} Nov 25 17:22:16 crc kubenswrapper[4802]: I1125 17:22:16.889924 4802 generic.go:334] "Generic (PLEG): container finished" podID="c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9" containerID="3015978f2f8c0bd32ed93b193e907125367ebb6850414bcf90966cecb06144a4" exitCode=0 Nov 25 17:22:16 crc kubenswrapper[4802]: I1125 17:22:16.889981 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfnrs" event={"ID":"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9","Type":"ContainerDied","Data":"3015978f2f8c0bd32ed93b193e907125367ebb6850414bcf90966cecb06144a4"} Nov 25 17:22:17 crc kubenswrapper[4802]: I1125 17:22:17.902540 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfnrs" event={"ID":"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9","Type":"ContainerStarted","Data":"aafb0eb80311dbb1dfd4229489a42ae981005812f852a121edb4f19305c80afb"} Nov 25 17:22:17 crc kubenswrapper[4802]: I1125 17:22:17.935953 4802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zfnrs" podStartSLOduration=2.484385177 podStartE2EDuration="4.935928986s" podCreationTimestamp="2025-11-25 17:22:13 +0000 UTC" firstStartedPulling="2025-11-25 17:22:14.871415882 +0000 UTC m=+2118.015763108" lastFinishedPulling="2025-11-25 17:22:17.322959721 +0000 UTC m=+2120.467306917" observedRunningTime="2025-11-25 17:22:17.931078025 +0000 UTC m=+2121.075425211" watchObservedRunningTime="2025-11-25 17:22:17.935928986 +0000 UTC m=+2121.080276192" Nov 25 17:22:19 crc kubenswrapper[4802]: I1125 17:22:19.377036 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xxb7x" Nov 25 17:22:19 crc kubenswrapper[4802]: I1125 17:22:19.377415 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xxb7x" Nov 25 17:22:19 crc kubenswrapper[4802]: I1125 17:22:19.431979 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xxb7x" Nov 25 17:22:19 crc kubenswrapper[4802]: I1125 17:22:19.972934 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xxb7x" Nov 25 17:22:20 crc kubenswrapper[4802]: I1125 17:22:20.582393 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xxb7x"] Nov 25 17:22:21 crc kubenswrapper[4802]: I1125 17:22:21.930659 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xxb7x" podUID="012112b2-13b7-42f2-b249-b443ca36abd8" containerName="registry-server" containerID="cri-o://82f089cd120da02f777e5947320755ed7736dcb2e09485c465859e2eb79ce353" gracePeriod=2 Nov 25 17:22:23 crc kubenswrapper[4802]: I1125 17:22:23.557683 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zfnrs" Nov 25 17:22:23 crc kubenswrapper[4802]: I1125 17:22:23.558285 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zfnrs" Nov 25 17:22:23 crc kubenswrapper[4802]: I1125 17:22:23.641473 4802 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zfnrs" Nov 25 17:22:23 crc kubenswrapper[4802]: I1125 17:22:23.988181 4802 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zfnrs" Nov 25 17:22:24 crc kubenswrapper[4802]: I1125 17:22:24.952046 4802 generic.go:334] "Generic (PLEG): container finished" podID="012112b2-13b7-42f2-b249-b443ca36abd8" containerID="82f089cd120da02f777e5947320755ed7736dcb2e09485c465859e2eb79ce353" exitCode=0 Nov 25 17:22:24 crc kubenswrapper[4802]: I1125 17:22:24.952139 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxb7x" event={"ID":"012112b2-13b7-42f2-b249-b443ca36abd8","Type":"ContainerDied","Data":"82f089cd120da02f777e5947320755ed7736dcb2e09485c465859e2eb79ce353"} Nov 25 17:22:24 crc kubenswrapper[4802]: I1125 17:22:24.952443 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxb7x" event={"ID":"012112b2-13b7-42f2-b249-b443ca36abd8","Type":"ContainerDied","Data":"f19a5f73cd1905546dca9e0aac63038940a17e6d2347f02dc271c75db1e50f4c"} Nov 25 17:22:24 crc kubenswrapper[4802]: I1125 17:22:24.952460 4802 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f19a5f73cd1905546dca9e0aac63038940a17e6d2347f02dc271c75db1e50f4c" Nov 25 17:22:24 crc kubenswrapper[4802]: I1125 17:22:24.953445 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xxb7x" Nov 25 17:22:25 crc kubenswrapper[4802]: I1125 17:22:25.147910 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/012112b2-13b7-42f2-b249-b443ca36abd8-utilities\") pod \"012112b2-13b7-42f2-b249-b443ca36abd8\" (UID: \"012112b2-13b7-42f2-b249-b443ca36abd8\") " Nov 25 17:22:25 crc kubenswrapper[4802]: I1125 17:22:25.151636 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/012112b2-13b7-42f2-b249-b443ca36abd8-utilities" (OuterVolumeSpecName: "utilities") pod "012112b2-13b7-42f2-b249-b443ca36abd8" (UID: "012112b2-13b7-42f2-b249-b443ca36abd8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:22:25 crc kubenswrapper[4802]: I1125 17:22:25.148116 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/012112b2-13b7-42f2-b249-b443ca36abd8-catalog-content\") pod \"012112b2-13b7-42f2-b249-b443ca36abd8\" (UID: \"012112b2-13b7-42f2-b249-b443ca36abd8\") " Nov 25 17:22:25 crc kubenswrapper[4802]: I1125 17:22:25.155680 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fjhk\" (UniqueName: \"kubernetes.io/projected/012112b2-13b7-42f2-b249-b443ca36abd8-kube-api-access-7fjhk\") pod \"012112b2-13b7-42f2-b249-b443ca36abd8\" (UID: \"012112b2-13b7-42f2-b249-b443ca36abd8\") " Nov 25 17:22:25 crc kubenswrapper[4802]: I1125 17:22:25.156345 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/012112b2-13b7-42f2-b249-b443ca36abd8-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 17:22:25 crc kubenswrapper[4802]: I1125 17:22:25.161507 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/012112b2-13b7-42f2-b249-b443ca36abd8-kube-api-access-7fjhk" (OuterVolumeSpecName: "kube-api-access-7fjhk") pod "012112b2-13b7-42f2-b249-b443ca36abd8" (UID: "012112b2-13b7-42f2-b249-b443ca36abd8"). InnerVolumeSpecName "kube-api-access-7fjhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:22:25 crc kubenswrapper[4802]: I1125 17:22:25.188759 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zfnrs"] Nov 25 17:22:25 crc kubenswrapper[4802]: I1125 17:22:25.221449 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/012112b2-13b7-42f2-b249-b443ca36abd8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "012112b2-13b7-42f2-b249-b443ca36abd8" (UID: "012112b2-13b7-42f2-b249-b443ca36abd8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:22:25 crc kubenswrapper[4802]: I1125 17:22:25.257438 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/012112b2-13b7-42f2-b249-b443ca36abd8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 17:22:25 crc kubenswrapper[4802]: I1125 17:22:25.257492 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fjhk\" (UniqueName: \"kubernetes.io/projected/012112b2-13b7-42f2-b249-b443ca36abd8-kube-api-access-7fjhk\") on node \"crc\" DevicePath \"\"" Nov 25 17:22:25 crc kubenswrapper[4802]: I1125 17:22:25.962409 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xxb7x" Nov 25 17:22:25 crc kubenswrapper[4802]: I1125 17:22:25.991875 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xxb7x"] Nov 25 17:22:25 crc kubenswrapper[4802]: I1125 17:22:25.998062 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xxb7x"] Nov 25 17:22:26 crc kubenswrapper[4802]: I1125 17:22:26.973157 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zfnrs" podUID="c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9" containerName="registry-server" containerID="cri-o://aafb0eb80311dbb1dfd4229489a42ae981005812f852a121edb4f19305c80afb" gracePeriod=2 Nov 25 17:22:27 crc kubenswrapper[4802]: I1125 17:22:27.470605 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zfnrs" Nov 25 17:22:27 crc kubenswrapper[4802]: I1125 17:22:27.486937 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnfld\" (UniqueName: \"kubernetes.io/projected/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9-kube-api-access-xnfld\") pod \"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9\" (UID: \"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9\") " Nov 25 17:22:27 crc kubenswrapper[4802]: I1125 17:22:27.487067 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9-catalog-content\") pod \"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9\" (UID: \"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9\") " Nov 25 17:22:27 crc kubenswrapper[4802]: I1125 17:22:27.487164 4802 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9-utilities\") pod \"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9\" (UID: \"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9\") " Nov 25 17:22:27 crc kubenswrapper[4802]: I1125 17:22:27.488537 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9-utilities" (OuterVolumeSpecName: "utilities") pod "c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9" (UID: "c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:22:27 crc kubenswrapper[4802]: I1125 17:22:27.495318 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9-kube-api-access-xnfld" (OuterVolumeSpecName: "kube-api-access-xnfld") pod "c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9" (UID: "c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9"). InnerVolumeSpecName "kube-api-access-xnfld". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 17:22:27 crc kubenswrapper[4802]: I1125 17:22:27.521582 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="012112b2-13b7-42f2-b249-b443ca36abd8" path="/var/lib/kubelet/pods/012112b2-13b7-42f2-b249-b443ca36abd8/volumes" Nov 25 17:22:27 crc kubenswrapper[4802]: I1125 17:22:27.585148 4802 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9" (UID: "c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 17:22:27 crc kubenswrapper[4802]: I1125 17:22:27.589192 4802 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 17:22:27 crc kubenswrapper[4802]: I1125 17:22:27.589447 4802 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 17:22:27 crc kubenswrapper[4802]: I1125 17:22:27.589530 4802 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnfld\" (UniqueName: \"kubernetes.io/projected/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9-kube-api-access-xnfld\") on node \"crc\" DevicePath \"\"" Nov 25 17:22:27 crc kubenswrapper[4802]: I1125 17:22:27.983481 4802 generic.go:334] "Generic (PLEG): container finished" podID="c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9" containerID="aafb0eb80311dbb1dfd4229489a42ae981005812f852a121edb4f19305c80afb" exitCode=0 Nov 25 17:22:27 crc kubenswrapper[4802]: I1125 17:22:27.983535 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfnrs" event={"ID":"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9","Type":"ContainerDied","Data":"aafb0eb80311dbb1dfd4229489a42ae981005812f852a121edb4f19305c80afb"} Nov 25 17:22:27 crc kubenswrapper[4802]: I1125 17:22:27.983546 4802 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zfnrs" Nov 25 17:22:27 crc kubenswrapper[4802]: I1125 17:22:27.983566 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfnrs" event={"ID":"c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9","Type":"ContainerDied","Data":"fbf99ba58a86850d12d2ba889a6d0b9311d25a0777920ff147ddf7c5a7bb23d0"} Nov 25 17:22:27 crc kubenswrapper[4802]: I1125 17:22:27.983587 4802 scope.go:117] "RemoveContainer" containerID="aafb0eb80311dbb1dfd4229489a42ae981005812f852a121edb4f19305c80afb" Nov 25 17:22:28 crc kubenswrapper[4802]: I1125 17:22:28.008299 4802 scope.go:117] "RemoveContainer" containerID="3015978f2f8c0bd32ed93b193e907125367ebb6850414bcf90966cecb06144a4" Nov 25 17:22:28 crc kubenswrapper[4802]: I1125 17:22:28.024324 4802 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zfnrs"] Nov 25 17:22:28 crc kubenswrapper[4802]: I1125 17:22:28.029319 4802 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zfnrs"] Nov 25 17:22:28 crc kubenswrapper[4802]: I1125 17:22:28.036822 4802 scope.go:117] "RemoveContainer" containerID="b01a93e161f49bd09c9973506ab5fe658bf6cb6f051521206b34253a4b3a90c4" Nov 25 17:22:28 crc kubenswrapper[4802]: I1125 17:22:28.055918 4802 scope.go:117] "RemoveContainer" containerID="aafb0eb80311dbb1dfd4229489a42ae981005812f852a121edb4f19305c80afb" Nov 25 17:22:28 crc kubenswrapper[4802]: E1125 17:22:28.056409 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aafb0eb80311dbb1dfd4229489a42ae981005812f852a121edb4f19305c80afb\": container with ID starting with aafb0eb80311dbb1dfd4229489a42ae981005812f852a121edb4f19305c80afb not found: ID does not exist" containerID="aafb0eb80311dbb1dfd4229489a42ae981005812f852a121edb4f19305c80afb" Nov 25 17:22:28 crc kubenswrapper[4802]: I1125 17:22:28.056435 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aafb0eb80311dbb1dfd4229489a42ae981005812f852a121edb4f19305c80afb"} err="failed to get container status \"aafb0eb80311dbb1dfd4229489a42ae981005812f852a121edb4f19305c80afb\": rpc error: code = NotFound desc = could not find container \"aafb0eb80311dbb1dfd4229489a42ae981005812f852a121edb4f19305c80afb\": container with ID starting with aafb0eb80311dbb1dfd4229489a42ae981005812f852a121edb4f19305c80afb not found: ID does not exist" Nov 25 17:22:28 crc kubenswrapper[4802]: I1125 17:22:28.056455 4802 scope.go:117] "RemoveContainer" containerID="3015978f2f8c0bd32ed93b193e907125367ebb6850414bcf90966cecb06144a4" Nov 25 17:22:28 crc kubenswrapper[4802]: E1125 17:22:28.056740 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3015978f2f8c0bd32ed93b193e907125367ebb6850414bcf90966cecb06144a4\": container with ID starting with 3015978f2f8c0bd32ed93b193e907125367ebb6850414bcf90966cecb06144a4 not found: ID does not exist" containerID="3015978f2f8c0bd32ed93b193e907125367ebb6850414bcf90966cecb06144a4" Nov 25 17:22:28 crc kubenswrapper[4802]: I1125 17:22:28.056753 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3015978f2f8c0bd32ed93b193e907125367ebb6850414bcf90966cecb06144a4"} err="failed to get container status \"3015978f2f8c0bd32ed93b193e907125367ebb6850414bcf90966cecb06144a4\": rpc error: code = NotFound desc = could not find container \"3015978f2f8c0bd32ed93b193e907125367ebb6850414bcf90966cecb06144a4\": container with ID starting with 3015978f2f8c0bd32ed93b193e907125367ebb6850414bcf90966cecb06144a4 not found: ID does not exist" Nov 25 17:22:28 crc kubenswrapper[4802]: I1125 17:22:28.056764 4802 scope.go:117] "RemoveContainer" containerID="b01a93e161f49bd09c9973506ab5fe658bf6cb6f051521206b34253a4b3a90c4" Nov 25 17:22:28 crc kubenswrapper[4802]: E1125 17:22:28.057040 4802 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b01a93e161f49bd09c9973506ab5fe658bf6cb6f051521206b34253a4b3a90c4\": container with ID starting with b01a93e161f49bd09c9973506ab5fe658bf6cb6f051521206b34253a4b3a90c4 not found: ID does not exist" containerID="b01a93e161f49bd09c9973506ab5fe658bf6cb6f051521206b34253a4b3a90c4" Nov 25 17:22:28 crc kubenswrapper[4802]: I1125 17:22:28.057054 4802 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b01a93e161f49bd09c9973506ab5fe658bf6cb6f051521206b34253a4b3a90c4"} err="failed to get container status \"b01a93e161f49bd09c9973506ab5fe658bf6cb6f051521206b34253a4b3a90c4\": rpc error: code = NotFound desc = could not find container \"b01a93e161f49bd09c9973506ab5fe658bf6cb6f051521206b34253a4b3a90c4\": container with ID starting with b01a93e161f49bd09c9973506ab5fe658bf6cb6f051521206b34253a4b3a90c4 not found: ID does not exist" Nov 25 17:22:29 crc kubenswrapper[4802]: I1125 17:22:29.516641 4802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9" path="/var/lib/kubelet/pods/c4ed3fe7-796b-48e8-9fe3-70d5598e4fc9/volumes" Nov 25 17:22:54 crc kubenswrapper[4802]: I1125 17:22:54.248908 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 17:22:54 crc kubenswrapper[4802]: I1125 17:22:54.249490 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 17:23:24 crc kubenswrapper[4802]: I1125 17:23:24.248018 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 17:23:24 crc kubenswrapper[4802]: I1125 17:23:24.248584 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 17:23:54 crc kubenswrapper[4802]: I1125 17:23:54.248668 4802 patch_prober.go:28] interesting pod/machine-config-daemon-h29wc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 17:23:54 crc kubenswrapper[4802]: I1125 17:23:54.249517 4802 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 17:23:54 crc kubenswrapper[4802]: I1125 17:23:54.249609 4802 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" Nov 25 17:23:54 crc kubenswrapper[4802]: I1125 17:23:54.250777 4802 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1535d557614337c68c32573c76f4ce74b2b995bdb4dc21eb0f632bc8caf7401a"} pod="openshift-machine-config-operator/machine-config-daemon-h29wc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 17:23:54 crc kubenswrapper[4802]: I1125 17:23:54.250881 4802 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerName="machine-config-daemon" containerID="cri-o://1535d557614337c68c32573c76f4ce74b2b995bdb4dc21eb0f632bc8caf7401a" gracePeriod=600 Nov 25 17:23:54 crc kubenswrapper[4802]: E1125 17:23:54.389003 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:23:54 crc kubenswrapper[4802]: I1125 17:23:54.731599 4802 generic.go:334] "Generic (PLEG): container finished" podID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" containerID="1535d557614337c68c32573c76f4ce74b2b995bdb4dc21eb0f632bc8caf7401a" exitCode=0 Nov 25 17:23:54 crc kubenswrapper[4802]: I1125 17:23:54.731662 4802 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" event={"ID":"4c3e23e7-2aae-4a81-b64f-4ed182a43d5e","Type":"ContainerDied","Data":"1535d557614337c68c32573c76f4ce74b2b995bdb4dc21eb0f632bc8caf7401a"} Nov 25 17:23:54 crc kubenswrapper[4802]: I1125 17:23:54.731711 4802 scope.go:117] "RemoveContainer" containerID="0ba8f23d451c7d83e7568a15af490bcf0f8abaf83f184136081d26b773dddc2e" Nov 25 17:23:54 crc kubenswrapper[4802]: I1125 17:23:54.732549 4802 scope.go:117] "RemoveContainer" containerID="1535d557614337c68c32573c76f4ce74b2b995bdb4dc21eb0f632bc8caf7401a" Nov 25 17:23:54 crc kubenswrapper[4802]: E1125 17:23:54.732932 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:24:04 crc kubenswrapper[4802]: I1125 17:24:04.787082 4802 scope.go:117] "RemoveContainer" containerID="94bd8b32081263dfca647b4dd016421c99a730aef8fe8a37386099669bfad830" Nov 25 17:24:04 crc kubenswrapper[4802]: I1125 17:24:04.813455 4802 scope.go:117] "RemoveContainer" containerID="2338638ca5ed88971ede29bb6724d04e0a9176a87cfa1aa9cade74a7f9a6b094" Nov 25 17:24:04 crc kubenswrapper[4802]: I1125 17:24:04.831269 4802 scope.go:117] "RemoveContainer" containerID="cc74930300789d532e334cd0c0a0af87dfbb754914e2bb5fad01711d93dbcc49" Nov 25 17:24:04 crc kubenswrapper[4802]: I1125 17:24:04.847092 4802 scope.go:117] "RemoveContainer" containerID="c1584998f2a38b0c604b0c04bc1dfcb59096d942d01df19b043c30b081576a85" Nov 25 17:24:04 crc kubenswrapper[4802]: I1125 17:24:04.890040 4802 scope.go:117] "RemoveContainer" containerID="dc84635203477ce7b1613a8ff8daa15a7f840dafa044be8e05162a1f1cb1bc0c" Nov 25 17:24:04 crc kubenswrapper[4802]: I1125 17:24:04.924646 4802 scope.go:117] "RemoveContainer" containerID="0fbf753bf1ccfb2ca11e00dd96a16d23da33d3c62c0e30d3844dd20435e8d72a" Nov 25 17:24:06 crc kubenswrapper[4802]: I1125 17:24:06.505534 4802 scope.go:117] "RemoveContainer" containerID="1535d557614337c68c32573c76f4ce74b2b995bdb4dc21eb0f632bc8caf7401a" Nov 25 17:24:06 crc kubenswrapper[4802]: E1125 17:24:06.506566 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:24:19 crc kubenswrapper[4802]: I1125 17:24:19.504559 4802 scope.go:117] "RemoveContainer" containerID="1535d557614337c68c32573c76f4ce74b2b995bdb4dc21eb0f632bc8caf7401a" Nov 25 17:24:19 crc kubenswrapper[4802]: E1125 17:24:19.505492 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:24:34 crc kubenswrapper[4802]: I1125 17:24:34.504994 4802 scope.go:117] "RemoveContainer" containerID="1535d557614337c68c32573c76f4ce74b2b995bdb4dc21eb0f632bc8caf7401a" Nov 25 17:24:34 crc kubenswrapper[4802]: E1125 17:24:34.505789 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:24:47 crc kubenswrapper[4802]: I1125 17:24:47.509911 4802 scope.go:117] "RemoveContainer" containerID="1535d557614337c68c32573c76f4ce74b2b995bdb4dc21eb0f632bc8caf7401a" Nov 25 17:24:47 crc kubenswrapper[4802]: E1125 17:24:47.510690 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:25:00 crc kubenswrapper[4802]: I1125 17:25:00.505353 4802 scope.go:117] "RemoveContainer" containerID="1535d557614337c68c32573c76f4ce74b2b995bdb4dc21eb0f632bc8caf7401a" Nov 25 17:25:00 crc kubenswrapper[4802]: E1125 17:25:00.506151 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" Nov 25 17:25:05 crc kubenswrapper[4802]: I1125 17:25:05.012171 4802 scope.go:117] "RemoveContainer" containerID="1a4940f0df0c4c4f6bb47145dfcfd12ee7b5dee23fd5ab9b7acdd0bd510c2ce8" Nov 25 17:25:13 crc kubenswrapper[4802]: I1125 17:25:13.505214 4802 scope.go:117] "RemoveContainer" containerID="1535d557614337c68c32573c76f4ce74b2b995bdb4dc21eb0f632bc8caf7401a" Nov 25 17:25:13 crc kubenswrapper[4802]: E1125 17:25:13.506021 4802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h29wc_openshift-machine-config-operator(4c3e23e7-2aae-4a81-b64f-4ed182a43d5e)\"" pod="openshift-machine-config-operator/machine-config-daemon-h29wc" podUID="4c3e23e7-2aae-4a81-b64f-4ed182a43d5e" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111363211024437 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111363212017355 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111356230016502 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111356230015452 5ustar corecore